Skip to content

Commit 73dbdf9

Browse files
Merge pull request #30 from PallHaraldsson/patch-1
Typo
2 parents d3aa078 + 3b5c0c7 commit 73dbdf9

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ summation algorithm (as implemented in Base.sum) starts losing accuracy as soon
8484
as the condition number increases, computing only noise when the condition
8585
number exceeds 1/ϵ≃10¹⁶. The same goes for the naive summation algorithm.
8686
In contrast, both compensated algorithms
87-
(Kahan-Babuska-Neumaier and Ogita-Rump-Oishi) still accurately compute the
87+
(KahanBabuskaNeumaier and OgitaRumpOishi) still accurately compute the
8888
result at this point, and start losing accuracy there, computing meaningless
8989
results when the condition nuber reaches 1/ϵ²≃10³². In effect these (simply)
9090
compensated algorithms produce the same results as if a naive summation had been
@@ -151,8 +151,8 @@ thousands of elements), the implementation is memory bound (as expected of a
151151
typical BLAS1 operation). Which is why we see significant decreases in the
152152
performance when the vector can’t fit into the L1, L2 or L3 cache.
153153

154-
On this AVX512-enabled system, the Kahan-Babuska-Neumaier implementation tends
155-
to be a little more efficient than the Ogita-Rump-Oishi algorithm (this would
154+
On this AVX512-enabled system, the KahanBabuskaNeumaier implementation tends
155+
to be a little more efficient than the OgitaRumpOishi algorithm (this would
156156
generally the opposite for AVX2 systems). When implemented with a suitable
157157
unrolling level and cache prefetching, these implementations are CPU-bound when
158158
vectors fit inside the L1 or L2 cache. However, when vectors are too large to

0 commit comments

Comments
 (0)