Starting a recipe (inference server) for a model
Once an AI model is available through a well known endpoint, it's easy to imagine a new world of applications that will connect and use the AI model. Podman AI Lab support AI applications as a set of containers that are connected together.
Podman AI Lab includes a Recipes Catalog that helps you navigate several core AI use cases and problem domains, such as chatbots, code generators, and text summarizers. Each recipe includes detailed explanations and sample applications that can be run with various large language models (LLMs). Experimenting with multiple models allows you to find the optimal one for your use case.
Prerequisites
Procedure
- Click the Podman AI Lab icon in the left navigation pane.
- In the Podman AI Lab navigation bar, click Recipe Catalog.
- Optional: Filter the recipe catalog based on the required tool, framework, or programming language.
- Click the More details link for the recipe that you want to start. The recipe summary page opens.
- Click Start. The Start recipe page opens.
- Select a model from the dropdown list.
- Click Start ChatBot recipe. This step might take some time to pull the recipe, copy the model to your Podman machine, start the inference server, load configurations, and create the application.
Verification
- Click Open details after the processing completes.
- View the created AI application.
- Click the Open AI App icon to view the running application in your browser.