Skip to content

Commit 5a2100c

Browse files
small fixes in documentation
1 parent be94364 commit 5a2100c

File tree

3 files changed

+18
-11
lines changed

3 files changed

+18
-11
lines changed

doc/source/cache_files.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,14 @@ Cache files
44
===========
55

66
A very useful feature of Kernel Tuner is the ability to store benchmarking results in a cache file during tuning. You can enable cache files by
7-
passing any filename to the ``cache=`` optional argument of ``tune_kernel()``.
7+
passing any filename to the ``cache=`` optional argument of ``tune_kernel``.
88

99
The benchmark results of individual kernel configurations are appended to the cache file as Kernel Tuner is running. This also allows Kernel Tuner
1010
to restart a ``tune_kernel()`` session from an existing cache file, should something have terminated the previous session before the run had
1111
completed. This happens quite often in HPC environments when a job reservation runs out.
1212

1313
Cache files enable a number of other features, such as simulations and visualizations. Simulations are useful for benchmarking optimization
14-
strategies. You can start a simulation by call tune_kernel with a cache file that contains the full search space and the ``simulation=True`` option.
14+
strategies. You can start a simulation by calling ``tune_kernel`` with a cache file that contains the full search space and the ``simulation=True`` option.
1515

1616
Cache files can be used to create visualizations of the search space. This even works while Kernel Tuner is still running. As the new results are
1717
coming, they are streamed to the visualization. Please see `Kernel Tuner Dashboard <https://github.com/KernelTuner/dashboard>`__.

doc/source/index.rst

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -98,4 +98,11 @@ If you use Kernel Tuner in research or research software, please cite the most r
9898
publisher={IEEE}
9999
}
100100

101-
101+
@article{schoonhoven2022going,
102+
author = {Schoonhoven, Richard and Veenboer, Bram, and van Werkhoven, Ben and Batenburg, K Joost},
103+
title = {Going green: optimizing GPUs for energy efficiency through model-steered auto-tuning},
104+
journal = {International Workshop on Performance Modeling, Benchmarking and Simulation
105+
of High Performance Computer Systems (PMBS) at Supercomputing (SC22)},
106+
year = {2022},
107+
url = {https://arxiv.org/abs/2211.07260}
108+
}

doc/source/quickstart.rst

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ Getting Started
33

44
So you have installed Kernel Tuner! That's great! But now you'd like to get started tuning some GPU code.
55

6-
Let's say we have a simple CUDA kernel stored in a file called vector_add_kernel.cu:
6+
Let's say we have a simple CUDA kernel stored in a file called ``vector_add_kernel.cu``:
77

88
.. code-block:: cuda
99
@@ -16,7 +16,7 @@ Let's say we have a simple CUDA kernel stored in a file called vector_add_kernel
1616
}
1717
1818
19-
This kernel simply performs a point-wise addition of vectors a and b and stores the result in c.
19+
This kernel simply performs a point-wise addition of vectors ``a`` and ``b`` and stores the result in ``c``.
2020

2121
To tune this kernel with Kernel Tuner, we are going to create the input and output data in Python using Numpy arrays.
2222

@@ -34,30 +34,30 @@ To tune this kernel with Kernel Tuner, we are going to create the input and outp
3434
To tell Kernel Tuner how it should call the kernel, we can create a list in Python that should correspond to
3535
our CUDA kernel's argument list with the same order and types.
3636

37-
.. code-block::python
37+
.. code-block:: python
3838
3939
args = [c, a, b, n]
4040
4141
So far, we have created the data structures needed by Kernel Tuner to call our kernel, but we have not yet specified what we
4242
want Kernel Tuner to tune in our kernel. For that, we create a dictionary that we call tune_params, in which keys correspond
4343
to tunable parameters in our kernel and the values are lists of values that these parameters may take.
4444

45-
.. code-block::python
45+
.. code-block:: python
4646
4747
tune_params = dict()
4848
tune_params["block_size_x"] = [32, 64, 128, 256, 512, 1024]
4949
50-
In the code above, we have inserted a key into our dictionary called "block_size_x". This is a special name for a tunable
50+
In the code above, we have inserted a key into our dictionary, namely ``"block_size_x"``. This is a special name for a tunable
5151
parameter that is recognized by Kernel Tuner to denote the size of our thread block in the x-dimension.
5252
For a full list of special parameter names, please see the :ref:`parameter-vocabulary`.
5353

54-
Alright, we are all set to start calling Kernel Tuner's main function, which is called tune_kernel.
54+
Alright, we are all set to start calling Kernel Tuner's main function, which is called ``tune_kernel``.
5555

56-
.. code-block::python
56+
.. code-block:: python
5757
5858
results, env = kernel_tuner.tune_kernel("vector_add", "vector_add_kernel.cu", size, args, tune_params)
5959
60-
In the above, tune_kernel takes five arguments:
60+
In the above, ``tune_kernel`` takes five arguments:
6161

6262
* The kernel name passed as a string
6363
* The filename of the kernel, also as a string

0 commit comments

Comments
 (0)