Skip to content

Commit e3eec95

Browse files
committed
Brought up to date with master
2 parents 4ddef4c + 359fe95 commit e3eec95

File tree

3 files changed

+5
-5
lines changed

3 files changed

+5
-5
lines changed

CONTRIBUTING.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ Before creating a pull request please ensure the following:
2828

2929
If you are in doubt on where to put your additions to the Kernel Tuner, please
3030
have look at the `design documentation
31-
<http://benvanwerkhoven.github.io/kernel_tuner/design.html>`__, or discuss it in the issue regarding your additions.
31+
<https://kerneltuner.github.io/kernel_tuner/stable/design.html>`__, or discuss it in the issue regarding your additions.
3232

3333
Development setup
3434
-----------------

doc/source/matrix_multiplication.ipynb

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -161,7 +161,7 @@
161161
"As we can see the execution times printed by `tune_kernel` already vary quite dramatically between the different values for `block_size_x` and `block_size_y`. However, even with the best thread block dimensions our kernel is still not very efficient.\n",
162162
"\n",
163163
"Therefore, we'll have a look at the Nvidia Visual Profiler to find that the utilization of our kernel is actually pretty low:\n",
164-
"![](https://raw.githubusercontent.com/kerneltuner/kernel_tuner/master/tutorial/matmul/matmul_naive.png)\n",
164+
"![](https://raw.githubusercontent.com/kerneltuner/kernel_tuner/master/doc/source/matmul/matmul_naive.png)\n",
165165
"There is however, a lot of opportunity for data reuse, which is realized by making the threads in a thread block collaborate."
166166
]
167167
},
@@ -270,7 +270,7 @@
270270
"source": [
271271
"This kernel drastically reduces memory bandwidth consumption. Compared to our naive kernel, it is about three times faster now, which comes from the highly increased memory utilization:\n",
272272
"\n",
273-
"![](https://raw.githubusercontent.com/kerneltuner/kernel_tuner/master/tutorial/matmul/matmul_shared.png)\n",
273+
"![](https://raw.githubusercontent.com/kerneltuner/kernel_tuner/master/doc/source/matmul/matmul_shared.png)\n",
274274
"\n",
275275
"The compute utilization has actually decreased slightly, which is due to the synchronization overhead, because ``__syncthread()`` is called frequently.\n",
276276
"\n",
@@ -422,7 +422,7 @@
422422
"source": [
423423
"As we can see the number of kernel configurations evaluated by the tuner has increased again. Also the performance has increased quite dramatically with roughly another factor 3. If we look at the Nvidia Visual Profiler output of our kernel we see the following:\n",
424424
"\n",
425-
"![](https://raw.githubusercontent.com/kerneltuner/kernel_tuner/master/tutorial/matmul/matmul.png)\n",
425+
"![](https://raw.githubusercontent.com/kerneltuner/kernel_tuner/master/doc/source/matmul/matmul.png)\n",
426426
"\n",
427427
"As expected, the compute utilization of our kernel has improved. There may even be some more room for improvement, but our tutorial on how to use Kernel Tuner ends here. In this tutorial, we have seen how you can use Kernel Tuner to tune kernels with a small number of tunable parameters, how to impose restrictions on the parameter space, and how to use grid divisor lists to specify how grid dimensions are computed."
428428
]

setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ def readme():
5050
'Topic :: System :: Distributed Computing',
5151
'Development Status :: 5 - Production/Stable',
5252
],
53-
install_requires=['numpy>=1.13.3', 'scipy>=1.8.1', 'jsonschema', 'python-constraint', 'xmltodict'],
53+
install_requires=['numpy>=1.13.3,<1.24.0', 'scipy>=1.8.1', 'jsonschema', 'python-constraint', 'xmltodict'],
5454
extras_require={
5555
'doc': ['sphinx', 'sphinx_rtd_theme', 'nbsphinx', 'pytest', 'ipython', 'markupsafe==2.0.1'],
5656
'cuda': ['pycuda', 'nvidia-ml-py', 'pynvml>=11.4.1'],

0 commit comments

Comments
 (0)