Skip to content

Commit ed54efc

Browse files
committed
Fix titles
1 parent 5626b9d commit ed54efc

1 file changed

Lines changed: 5 additions & 5 deletions

File tree

src/components/fundable/descriptions/Float16SupportInXsimd.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
1-
# FP16 Support in xsimd
1+
#### Overview
22

3-
xsimd is a C++ header-only library that abstracts SIMD (vectorization) intrinsics behind a single, generic API.
3+
Xsimd is a C++ header-only library that abstracts SIMD (vectorization) intrinsics behind a single, generic API.
44
The same code — `xsimd::batch<float>` — compiles to optimal machine code on x86 SSE/AVX, ARM NEON SVE, RISC-V, and WebAssembly, with no runtime overhead.
55
When an intrinsic is missing on a given target, xsimd falls back gracefully rather than failing or leaving the developer to write platform-specific branches.
66
This is why projects like Mozilla Firefox, Apache Arrow, Meta Velox, KDE Krita, and Pythran have adopted it as their vectorization layer.
@@ -13,7 +13,7 @@ xsimd currently has no FP16 support, forcing its users to drop out of the generi
1313

1414
We propose to add vectorized FP16 support to xsimd — native FP16 operations where hardware supports them, and correct fallbacks elsewhere.
1515

16-
## Why FP16 Matters
16+
#### Why FP16 Matters
1717

1818
**Memory bandwidth is a bottleneck.** Modern CPUs and GPUs are not compute-bound — they are memory-bandwidth-bound.
1919
FP16 cuts data size in half versus FP32.
@@ -37,7 +37,7 @@ FP16 conversion and arithmetic are now widely available across all major SIMD fa
3737
This affects NEON operations on modern smartphones and all Apple silicon M-chips.
3838
Coverage is extended server side with both SVE and SVE2 supporting FP16.
3939

40-
## Proposed Work
40+
#### Proposed Work
4141

4242
This proposal covers foundational FP16 support: native FP16 operations on platforms that provide hardware acceleration, and correct, efficient fallbacks everywhere else.
4343

@@ -46,7 +46,7 @@ Concretely, this means:
4646
- Support for converting from and to `batch<float>`, mapping to the optimal hardware instruction where available, and a correct SIMD algorithm elsewhere.
4747
- Native FP16 arithmetic operations — add, multiply, FMA, min, max, and comparison — on backends that provide hardware support, with FP32-based fallbacks on those that do not
4848

49-
## Impact
49+
#### Impact
5050

5151
Funding this development will directly open xsimd to the rapidly growing landscape of LLM and machine
5252
learning workflows: local inference engines, model weight processing, and embedding pipelines.

0 commit comments

Comments
 (0)