Skip to content

Commit

Permalink
Update Sphinx documentation, commit 33d7f6e [skip ci].
Browse files Browse the repository at this point in the history
  • Loading branch information
bluescarni committed Jun 11, 2024
1 parent 16dfd35 commit 6c50586
Show file tree
Hide file tree
Showing 5 changed files with 13 additions and 7 deletions.
2 changes: 2 additions & 0 deletions _sources/changelog.rst
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,8 @@ New
- Add thermonets: neural, differentiable, high-performance
models for the Earth's thermosphere density
(`#425 <https://github.com/bluescarni/heyoka/pull/425>`__).
- Add support for variational ODE systems and Taylor map computation
(`#422 <https://github.com/bluescarni/heyoka/pull/422>`__).
- Introduce the ability to represent unknown functions
and their derivatives in the expression system
(`#416 <https://github.com/bluescarni/heyoka/pull/416>`__).
Expand Down
7 changes: 4 additions & 3 deletions _sources/tut_batch_mode.rst
Original file line number Diff line number Diff line change
Expand Up @@ -18,13 +18,14 @@ batch mode can lead to an increase in floating-point throughput up to a factor o
It is important to emphasise that batch mode does not reduce
the CPU time required to integrate a system of ODEs. Rather, as a fine-grained
form of data parallelism, batch mode allows to integrate multiple ODE systems in parallel
at no additional cost, and it is thus most useful when the need arise
at (almost) no additional cost, and it is thus most useful when the need arise
to integrate the same ODE system with different initial conditions and parameters.

Although batch mode can in principle be used with all floating-point types supported
Although batch mode can in principle be used with all the fundamental C++ floating-point types supported
by heyoka, in practice at this time no CPU provides SIMD instructions for extended-precision
datatypes. Thus, here we will consider the application of batch mode only to
standard ``double`` precision computations.
standard ``double`` precision computations (keeping in mind that batch mode is also supported
for single-precision ``float`` computations).

The value of the batch size :math:`n` can be freely chosen by the user. In order
to achieve optimal performance, however, :math:`n` should match the SIMD width of the
Expand Down
2 changes: 2 additions & 0 deletions changelog.html
Original file line number Diff line number Diff line change
Expand Up @@ -594,6 +594,8 @@ <h3>New<a class="headerlink" href="#new" title="Link to this heading">#</a></h3>
<li><p>Add thermonets: neural, differentiable, high-performance
models for the Earth’s thermosphere density
(<a class="reference external" href="https://github.com/bluescarni/heyoka/pull/425">#425</a>).</p></li>
<li><p>Add support for variational ODE systems and Taylor map computation
(<a class="reference external" href="https://github.com/bluescarni/heyoka/pull/422">#422</a>).</p></li>
<li><p>Introduce the ability to represent unknown functions
and their derivatives in the expression system
(<a class="reference external" href="https://github.com/bluescarni/heyoka/pull/416">#416</a>).</p></li>
Expand Down
2 changes: 1 addition & 1 deletion searchindex.js

Large diffs are not rendered by default.

7 changes: 4 additions & 3 deletions tut_batch_mode.html
Original file line number Diff line number Diff line change
Expand Up @@ -434,12 +434,13 @@ <h2> Contents </h2>
<p>It is important to emphasise that batch mode does not reduce
the CPU time required to integrate a system of ODEs. Rather, as a fine-grained
form of data parallelism, batch mode allows to integrate multiple ODE systems in parallel
at no additional cost, and it is thus most useful when the need arise
at (almost) no additional cost, and it is thus most useful when the need arise
to integrate the same ODE system with different initial conditions and parameters.</p>
<p>Although batch mode can in principle be used with all floating-point types supported
<p>Although batch mode can in principle be used with all the fundamental C++ floating-point types supported
by heyoka, in practice at this time no CPU provides SIMD instructions for extended-precision
datatypes. Thus, here we will consider the application of batch mode only to
standard <code class="docutils literal notranslate"><span class="pre">double</span></code> precision computations.</p>
standard <code class="docutils literal notranslate"><span class="pre">double</span></code> precision computations (keeping in mind that batch mode is also supported
for single-precision <code class="docutils literal notranslate"><span class="pre">float</span></code> computations).</p>
<p>The value of the batch size <span class="math notranslate nohighlight">\(n\)</span> can be freely chosen by the user. In order
to achieve optimal performance, however, <span class="math notranslate nohighlight">\(n\)</span> should match the SIMD width of the
processor in use. Because at this time the most widespread SIMD instruction set is
Expand Down

0 comments on commit 6c50586

Please sign in to comment.