diff --git a/docs/development.md b/docs/development.md index 2baece56..42e62563 100644 --- a/docs/development.md +++ b/docs/development.md @@ -110,26 +110,15 @@ Example [.env](../.env.example) file. ## Local Dev Chat Environment -### 1) Using the ilab command line tool +### Using the ilab command line tool -For the chat functionality to work you need a ilab model chat instance. To run this locally: - -`cd server` - -[https://github.com/instructlab/instructlab?tab=readme-ov-file#-getting-started](https://github.com/instructlab/instructlab?tab=readme-ov-file#-getting-started) +For the chat functionality to work you need a ilab model chat instance. To run this locally see [Ilab Getting Started](https://github.com/instructlab/instructlab?tab=readme-ov-file#-getting-started). After you use the `ilab serve` command you should have, by default, a chat server instance running on port 8000. -### 2) Using Podman - -#### Current issues - -- The docker image that runs the server does not utilise Mac Metal GPU and therefore is very slow when answering prompts -- The docker image is very large as it contains the model itself. Potential to have the model incorporated via a docker volume to reduce the size of the actual image. - -`docker run -p 8000:8000 aevo987654/instructlab_chat_8000:v2` +### 2) Running in a Container -This should run a server on port 8000 +See the upstream CLI documentation for running ilab in a [container](https://github.com/instructlab/instructlab/tree/main/containers). ### Configuring the chat environment to use a local ilab model chat instance diff --git a/package-lock.json b/package-lock.json index d7870a22..edb7d18e 100644 --- a/package-lock.json +++ b/package-lock.json @@ -5430,12 +5430,6 @@ "uuid": "dist/bin/uuid" } }, - "node_modules/next/node_modules/@next/env": { - "version": "15.2.0", - "resolved": "https://registry.npmjs.org/@next/env/-/env-15.2.0.tgz", - "integrity": "sha512-eMgJu1RBXxxqqnuRJQh5RozhskoNUDHBFybvi+Z+yK9qzKeG7dadhv/Vp1YooSZmCnegf7JxWuapV77necLZNA==", - "license": "MIT" - }, "node_modules/next/node_modules/postcss": { "version": "8.4.31", "funding": [ diff --git a/server/Containerfile b/server/Containerfile deleted file mode 100644 index 18da2bc7..00000000 --- a/server/Containerfile +++ /dev/null @@ -1,21 +0,0 @@ -FROM python:3.11 - -# Set working directory -WORKDIR /app - -RUN pip install --upgrade pip -RUN pip install --no-cache-dir instructlab==0.16.1 - -# Copy project files to the working directory -COPY config.yaml . - -# Download the merlinite model -RUN ilab download - -# Copy project files to the working directory -COPY . . - -EXPOSE 8000 - -# Run the chat server with the specified model family and model file -CMD ["ilab", "serve", "--model-family", "merlinite", "--model-path", "models/merlinite-7b-lab-Q4_K_M.gguf"] \ No newline at end of file diff --git a/server/config.yaml b/server/config.yaml deleted file mode 100644 index ee4f316b..00000000 --- a/server/config.yaml +++ /dev/null @@ -1,26 +0,0 @@ -chat: - context: default - greedy_mode: false - logs_dir: data/chatlogs - max_tokens: null - model: models/merlinite-7b-lab-Q4_K_M.gguf - session: null - vi_mode: false - visible_overflow: true -general: - log_level: INFO -generate: - chunk_word_count: 1000 - model: models/merlinite-7b-lab-Q4_K_M.gguf - num_cpus: 10 - num_instructions: 100 - output_dir: generated - prompt_file: prompt.txt - seed_file: seed_tasks.json - taxonomy_base: origin/main - taxonomy_path: taxonomy -serve: - gpu_layers: -1 - host_port: 0.0.0.0:8000 - max_ctx_size: 4096 - model_path: models/merlinite-7b-lab-Q4_K_M.gguf