Skip to content

rh-ecosystem-edge/assisted-chat

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Assisted Chat

This repo builds the images for and runs in a podman pod the following services:

  • assisted-service-mcp
  • lightspeed-core + llama-stack (Same container, built separately)
  • assisted UI
  • inspector
  • postgres

Pre-requisites

  • the ocm command (get it here)

  • Either Vertex AI service account credentials or a Gemini API key. If using a Gemini API key, run make generate for instructions on how to get one.

  • An OpenShift pull-secret (https://cloud.redhat.com/openshift/install/pull-secret)

    • Apply it to podman by copying it to ~/.config/containers/auth.json
  • Some report needing a Red Hat Developer Subscription (required for building container images)

    • You might not need one
    • Sign up for free at https://developers.redhat.com/register
      1. Login to container registry: podman login registry.access.redhat.com
      2. Register your system: sudo subscription-manager register --username <your-username>
    • Verify registration: subscription-manager status

Dependencies

dnf install jq fzf

pip install yq uv

Submodules

git submodule init
git submodule update

Overview

Basics

This repo uses a podman pod to run all the required services for the Assisted chat bot. The pod is defined in assisted-chat-pod.yaml. The pod images are built from the git submodules in this repo. We use a Makefile for running commands.

You are expected to first run make generate to generate the .env and configuration files. make generate is interactive and will ask you whether you want to use a Gemini API key or Vertex AI service account credentials.

Before running the pod, you should run make build-images to build all the required images. You can also build individual images using make build-* commands (see make help).

Once you've generated the configuration files and built the images, you can run the pod with make run. This will start the pod and follow its logs. You can safely ctrl+c out of it and the pod will continue running in the background. To stop the pod, run make stop and optionally make rm to remove it completely. If you wish to see the logs again while it's running, use make logs.

To interact with assisted-chat, use make query.

Extra

You can also use make query-int and make query-stage to interact with the SaaS deployed integration and staging environments, respectively. This does not require the pod to be running.

You can interact with the model directly with mcphost through make mcphost

  • it's a bit weird, you have to first Ctrl+C out of it, then quickly run make mcphost again for it to actually work. Please fix it if you figure out why it's happening.

To see the contents of the postgres database, use make psql. This will put you in a psql shell. The database is used by both llama-stack and lightspeed-stack. llama-stack uses the default schema while lightspeed-stack uses the lightspeed-stack schema. The psql shell is pre-configured to see both schemas. If you're running lightspeed-stack without a database configuration you can use make sqlite to browse the contents of the temporary SQLite database.

Override

  • You can set LIGHTSPEED_STACK_IMAGE_OVERRIDE in .env to your own lightspeed-stack image (e.g. quay.io/lightspeed-core/lightspeed-stack:latest) to replace the locally built one used in the pod

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 14