Skip to content

Commit dcc21a6

Browse files
committed
OpenInference tutorial
1 parent 3f9cf78 commit dcc21a6

14 files changed

Lines changed: 658 additions & 12 deletions

File tree

tutorials/docs/etl/expose.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -109,6 +109,6 @@ Go to the Kubernetes Resource Manager component (available from dashboard) and g
109109

110110
The platform by default support exposing the methods at the subdomains of ``services.<platform-domain>``, where platform-domain is the domain of the platform instance.
111111

112-
![KRM APIGW image](../../images/apigw-krm.png)
112+
![KRM APIGW image](../images/apigw-krm.png)
113113

114114
*Save* and, after a few moments, you will be able to call the API at the address you defined! If you set *Authentication* to *Basic*, don't forget that you have to provide the credentials.

tutorials/docs/etl/pipeline.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,6 +58,6 @@ workflow.run("pipeline", parameters={"url": di.key}, wait=True)
5858

5959
It is possible to monitor the execution in the Core console:
6060

61-
![Pipeline image](./images/pipeline.png)
61+
![Pipeline image](../images/pipeline.png)
6262

6363
The next section will describe how to expose this newly obtained dataset as a REST API.

tutorials/docs/flower/flower.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ The platform support this approach natively integrating the Flower framework in
99
- support creating a federation, with central Superlink node and a set of client Supernodes distributed potentially outside of the platform in a secure manner (with TLS verification and client authentication)
1010
- activate the training procedures defined with the server coordination code and client training code managed by the platform.
1111

12-
See more details in the description of the corresponding [Flower runtime](../../runtimes/fl.md).
12+
See more details in the description of the corresponding [Flower runtime](../../../runtimes/fl/).
1313

1414
This tutorial demonstrates how to use the Flower FL framework for execution of federated learning tasks. The tutorial is based on official Pandas example of Flower framework.
1515

tutorials/docs/images/pipeline.png

67.6 KB
Loading

tutorials/docs/ml/deploy.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -130,6 +130,6 @@ Go to the Kubernetes Resource Manager component (available from dashboard) and g
130130

131131
The platform by default support exposing the methods at the subdomains of ``services.<platform-domain>``, where platform-domain is the domain of the platform instance.
132132

133-
![KRM APIGW image](../../images/apigw-krm.png)
133+
![KRM APIGW image](../images/apigw-krm.png)
134134

135135
*Save* and, after a few moments, you will be able to call the API at the address you defined! If you set *Authentication* to *Basic*, don't forget that you have to provide the credentials.

tutorials/docs/mlllm/llm.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ Serve the model:
4949
llm_run = llm_function.run(action="serve", profile="1xa100", wait=True)
5050
```
5151

52-
Please note the use of the ``profile`` parameter. As the LLM models require specific hardware (GPU in particular), it is necessary to specify the HW requirements as described in the [Configuring Kubernetes executions](../../tasks/kubernetes-resources.md) section. In particular, it is possible to rely on the predefined resource templates of the platform deployment.
52+
Please note the use of the ``profile`` parameter. As the LLM models require specific hardware (GPU in particular), it is necessary to specify the HW requirements as described in the [Configuring Kubernetes executions](../../../tasks/kubernetes-resources/) section. In particular, it is possible to rely on the predefined resource templates of the platform deployment.
5353

5454
As in other scenarios, you need to wait a bit for the service to become available.
5555
Once the service becomes available, it is possible to make the calls:
@@ -133,7 +133,7 @@ llm_run = llm_function.run(action="serve",
133133
in some cases up to 10 minutes.
134134

135135
As in case of classification models, the LLM models require specific hardware (GPU in particular), it is necessary
136-
to specify the HW requirements as described in the [Configuring Kubernetes executions](../../tasks/kubernetes-resources.md) section. In particular, it is possible to rely on the predefined resource templates of the platform deployment.
136+
to specify the HW requirements as described in the [Configuring Kubernetes executions](../../../tasks/kubernetes-resources/) section. In particular, it is possible to rely on the predefined resource templates of the platform deployment.
137137

138138
Once the service becomes available, it is possible to make the calls. For example, for the completion requests:
139139

@@ -358,7 +358,7 @@ llm_run = llm_function.run(action="serve", profile="1xa100", volumes=[{
358358
```
359359

360360
Please note the use of the ``profile`` parameter. As the LLM models require specific hardware (GPU in particular), it is necessary
361-
to specify the HW requirements as described in the [Configuring Kubernetes executions](../../tasks/kubernetes-resources.md) section. In particular, it is possible to rely on the predefined resource templates of the platform deployment. Also, in case of large models the default disk space may be insufficient and an extra volume should be configured for the underlying deployment.
361+
to specify the HW requirements as described in the [Configuring Kubernetes executions](../../../tasks/kubernetes-resources) section. In particular, it is possible to rely on the predefined resource templates of the platform deployment. Also, in case of large models the default disk space may be insufficient and an extra volume should be configured for the underlying deployment.
362362

363363
Once the service becomes available, it is possible to make the calls:
364364

tutorials/docs/mlllm/llmkubeai.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ For what concerns LLM tasks, currently KubeAI runtime allows for deploying the m
1010

1111
To accomplish this, it is possible to use one of the KubeAI-supported runtimes, namely [vLLM](https://docs.vllm.ai/), [OLlama](https://ollama.com/), and [Infinity](https://michaelfeil.eu/infinity). in case of vLLM also adapters are supported.
1212

13-
For details about the specification, see the corresponding section of [Modelserve](../../runtimes/modelserve.md) reference.
13+
For details about the specification, see the corresponding section of [Modelserve](../../../runtimes/modelserve) reference.
1414

1515
## Exposing Text Generation Models
1616

tutorials/docs/mlmlflow/deploy.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,6 +60,6 @@ Go to the Kubernetes Resource Manager component (available from dashboard) and g
6060

6161
The platform by default support exposing the methods at the subdomains of ``services.<platform-domain>``, where platform-domain is the domain of the platform instance.
6262

63-
![KRM APIGW image](../../images/apigw-krm.png)
63+
![KRM APIGW image](../images/apigw-krm.png)
6464

6565
*Save* and, after a few moments, you will be able to call the API at the address you defined! If you set *Authentication* to *Basic*, don't forget that you have to provide the credentials.

tutorials/docs/mlsklearn/deploy.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,6 @@ Go to the Kubernetes Resource Manager component (available from dashboard) and g
4747

4848
The platform by default support exposing the methods at the subdomains of ``services.<platform-domain>``, where platform-domain is the domain of the platform instance.
4949

50-
![KRM APIGW image](../../images/apigw-krm.png)
50+
![KRM APIGW image](../images/apigw-krm.png)
5151

5252
*Save* and, after a few moments, you will be able to call the API at the address you defined! If you set *Authentication* to *Basic*, don't forget that you have to provide the credentials.

0 commit comments

Comments
 (0)