Skip to content

Commit 0455f30

Browse files
committed
Update Sphinx documentation
1 parent 1ac47fb commit 0455f30

File tree

6 files changed

+74
-58
lines changed

6 files changed

+74
-58
lines changed

.doctrees/configuration-guide.doctree

1.5 KB
Binary file not shown.
0 Bytes
Binary file not shown.

.doctrees/environment.pickle

0 Bytes
Binary file not shown.

_sources/configuration-guide.rst.txt

Lines changed: 36 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,7 @@ Configuration Guide
22
####################
33

44
The following guide will help you configure ``values.yaml`` file for a SuperSONIC deployment.
5+
The full list of parameters can be found in the `Configuration Reference <configuration-reference>`_.
56

67
Triton Inference Server Configuration
78
****************************************
@@ -19,6 +20,8 @@ Triton version must be specified in the ``triton.image`` parameter in the values
1920
2. Configure Triton model repository.
2021
=============================================
2122

23+
- To learn about the structure of model repositories, refer to the
24+
`NVIDIA Model Repository Guide <https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/user_guide/model_repository.html>`_.
2225
- Model repositories are specified in the ``triton.args`` parameter in the values file.
2326
The parameter contains the full command that launches a Triton server; you can specify
2427
one or multiple model repositories via the ``--model-repository`` flag.
@@ -41,39 +44,46 @@ Triton version must be specified in the ``triton.image`` parameter in the values
4144
- Make sure that the model repository paths exist. You can load models from a volume mounted to the Triton container.
4245
The following options for model repository mouning are provided via ``triton.modelRepository`` parameter in ``values.yaml``:
4346

44-
.. code-block:: yaml
47+
.. raw:: html
4548

46-
# -- Model repository configuration
47-
modelRepository:
48-
# Set to `true` to enable model repository mounting
49-
enabled: true
49+
<details>
50+
<summary>Model repository configuration</summary>
51+
52+
.. code-block:: yaml
53+
54+
# -- Model repository configuration
55+
modelRepository:
56+
# Set to `true` to enable model repository mounting
57+
enabled: true
58+
59+
# -- Model repository mount path (e.g /cvmfs/)
60+
mountPath: ""
5061

51-
# -- Model repository mount path (e.g /cvmfs/)
52-
mountPath: ""
62+
## Model repository options:
5363

54-
## Model repository options:
64+
## Option 1: mount an arbitrary PersistentVolumeClaim
65+
storageType: "pvc"
66+
pvc:
67+
claimName:
5568

56-
## Option 1: mount an arbitrary PersistentVolumeClaim
57-
storageType: "pvc"
58-
pvc:
59-
claimName:
69+
## -- OR --
70+
## Option 2: mount CVMFS as PersistentVolumeClaim (CVMFS StorageClass must be installed at the cluster)
71+
storageType: "cvmfs-pvc"
72+
73+
## -- OR --
74+
## Option 3: mount CVMFS via hostPath (CVMFS must be already mounted on the nodes)
75+
storageType: "cvmfs"
6076

61-
## -- OR --
62-
## Option 2: mount CVMFS as PersistentVolumeClaim (CVMFS StorageClass must be installed at the cluster)
63-
storageType: "cvmfs-pvc"
64-
65-
## -- OR --
66-
## Option 3: mount CVMFS via hostPath (CVMFS must be already mounted on the nodes)
67-
storageType: "cvmfs"
77+
## -- OR --
78+
## Option 4: mount an NFS storage volume
79+
storageType: "nfs"
80+
nfs:
81+
server:
82+
path:
6883

69-
## -- OR --
70-
## Option 4: mount an NFS storage volume
71-
storageType: "nfs"
72-
nfs:
73-
server:
74-
path:
84+
</details>
7585

76-
1. Select resources for Triton pods.
86+
3. Select resources for Triton pods.
7787
=============================================
7888

7989
- You can configure CPU, memory, and GPU resources for Triton pods via the ``triton.resources`` parameter in the values file.

configuration-guide.html

Lines changed: 37 additions & 31 deletions
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,8 @@
8484

8585
<section id="configuration-guide">
8686
<h1>Configuration Guide<a class="headerlink" href="#configuration-guide" title="Link to this heading"></a></h1>
87-
<p>The following guide will help you configure <code class="docutils literal notranslate"><span class="pre">values.yaml</span></code> file for a SuperSONIC deployment.</p>
87+
<p>The following guide will help you configure <code class="docutils literal notranslate"><span class="pre">values.yaml</span></code> file for a SuperSONIC deployment.
88+
The full list of parameters can be found in the <a class="reference external" href="configuration-reference">Configuration Reference</a>.</p>
8889
<section id="triton-inference-server-configuration">
8990
<h2>Triton Inference Server Configuration<a class="headerlink" href="#triton-inference-server-configuration" title="Link to this heading"></a></h2>
9091
<section id="select-a-triton-inference-server-version">
@@ -100,6 +101,8 @@ <h3>1. Select a Triton Inference Server version<a class="headerlink" href="#sele
100101
<section id="configure-triton-model-repository">
101102
<h3>2. Configure Triton model repository.<a class="headerlink" href="#configure-triton-model-repository" title="Link to this heading"></a></h3>
102103
<ul>
104+
<li><p>To learn about the structure of model repositories, refer to the
105+
<a class="reference external" href="https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/user_guide/model_repository.html">NVIDIA Model Repository Guide</a>.</p></li>
103106
<li><p>Model repositories are specified in the <code class="docutils literal notranslate"><span class="pre">triton.args</span></code> parameter in the values file.
104107
The parameter contains the full command that launches a Triton server; you can specify
105108
one or multiple model repositories via the <code class="docutils literal notranslate"><span class="pre">--model-repository</span></code> flag.</p></li>
@@ -119,43 +122,46 @@ <h3>2. Configure Triton model repository.<a class="headerlink" href="#configure-
119122
</div>
120123
</li>
121124
<li><p>Make sure that the model repository paths exist. You can load models from a volume mounted to the Triton container.
122-
The following options for model repository mouning are provided via <code class="docutils literal notranslate"><span class="pre">triton.modelRepository</span></code> parameter in <code class="docutils literal notranslate"><span class="pre">values.yaml</span></code>:</p>
123-
<div class="highlight-yaml notranslate"><div class="highlight"><pre><span></span><span class="c1"># -- Model repository configuration</span>
124-
<span class="nt">modelRepository</span><span class="p">:</span>
125-
<span class="w"> </span><span class="c1"># Set to `true` to enable model repository mounting</span>
126-
<span class="w"> </span><span class="nt">enabled</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">true</span>
125+
The following options for model repository mouning are provided via <code class="docutils literal notranslate"><span class="pre">triton.modelRepository</span></code> parameter in <code class="docutils literal notranslate"><span class="pre">values.yaml</span></code>:</p></li>
126+
</ul>
127+
<details>
128+
<summary>Model repository configuration</summary>
127129

128-
<span class="w"> </span><span class="c1"># -- Model repository mount path (e.g /cvmfs/)</span>
129-
<span class="w"> </span><span class="nt">mountPath</span><span class="p">:</span><span class="w"> </span><span class="s">&quot;&quot;</span>
130+
.. code-block:: yaml
130131

131-
<span class="w"> </span><span class="c1">## Model repository options:</span>
132+
# -- Model repository configuration
133+
modelRepository:
134+
# Set to `true` to enable model repository mounting
135+
enabled: true
132136

133-
<span class="w"> </span><span class="c1">## Option 1: mount an arbitrary PersistentVolumeClaim</span>
134-
<span class="w"> </span><span class="nt">storageType</span><span class="p">:</span><span class="w"> </span><span class="s">&quot;pvc&quot;</span>
135-
<span class="w"> </span><span class="nt">pvc</span><span class="p">:</span>
136-
<span class="w"> </span><span class="nt">claimName</span><span class="p">:</span>
137+
# -- Model repository mount path (e.g /cvmfs/)
138+
mountPath: ""
137139

138-
<span class="w"> </span><span class="c1">## -- OR --</span>
139-
<span class="w"> </span><span class="c1">## Option 2: mount CVMFS as PersistentVolumeClaim (CVMFS StorageClass must be installed at the cluster)</span>
140-
<span class="w"> </span><span class="nt">storageType</span><span class="p">:</span><span class="w"> </span><span class="s">&quot;cvmfs-pvc&quot;</span>
140+
## Model repository options:
141141

142-
<span class="w"> </span><span class="c1">## -- OR --</span>
143-
<span class="w"> </span><span class="c1">## Option 3: mount CVMFS via hostPath (CVMFS must be already mounted on the nodes)</span>
144-
<span class="w"> </span><span class="nt">storageType</span><span class="p">:</span><span class="w"> </span><span class="s">&quot;cvmfs&quot;</span>
142+
## Option 1: mount an arbitrary PersistentVolumeClaim
143+
storageType: "pvc"
144+
pvc:
145+
claimName:
145146

146-
<span class="w"> </span><span class="c1">## -- OR --</span>
147-
<span class="w"> </span><span class="c1">## Option 4: mount an NFS storage volume</span>
148-
<span class="w"> </span><span class="nt">storageType</span><span class="p">:</span><span class="w"> </span><span class="s">&quot;nfs&quot;</span>
149-
<span class="w"> </span><span class="nt">nfs</span><span class="p">:</span>
150-
<span class="w"> </span><span class="nt">server</span><span class="p">:</span>
151-
<span class="w"> </span><span class="nt">path</span><span class="p">:</span>
152-
</pre></div>
153-
</div>
154-
</li>
155-
</ul>
156-
</section>
147+
## -- OR --
148+
## Option 2: mount CVMFS as PersistentVolumeClaim (CVMFS StorageClass must be installed at the cluster)
149+
storageType: "cvmfs-pvc"
150+
151+
## -- OR --
152+
## Option 3: mount CVMFS via hostPath (CVMFS must be already mounted on the nodes)
153+
storageType: "cvmfs"
154+
155+
## -- OR --
156+
## Option 4: mount an NFS storage volume
157+
storageType: "nfs"
158+
nfs:
159+
server:
160+
path:
161+
162+
</details></section>
157163
<section id="select-resources-for-triton-pods">
158-
<h3>1. Select resources for Triton pods.<a class="headerlink" href="#select-resources-for-triton-pods" title="Link to this heading"></a></h3>
164+
<h3>3. Select resources for Triton pods.<a class="headerlink" href="#select-resources-for-triton-pods" title="Link to this heading"></a></h3>
159165
<ul>
160166
<li><p>You can configure CPU, memory, and GPU resources for Triton pods via the <code class="docutils literal notranslate"><span class="pre">triton.resources</span></code> parameter in the values file.</p>
161167
<div class="highlight-yaml notranslate"><div class="highlight"><pre><span></span><span class="c1"># Example:</span>

0 commit comments

Comments
 (0)