From 401520f8a07d62d6538e984a7147e3a403bb00f3 Mon Sep 17 00:00:00 2001 From: Bennet Date: Fri, 5 Apr 2024 00:05:46 +0200 Subject: [PATCH 1/6] Update release.yml --- .github/workflows/release.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index 41b6f65..b55f009 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -226,7 +226,7 @@ jobs: - name: Install dependencies run: poetry install --only test --no-root --no-cache - name: Run pytest - run: ZEEBE_TEST_IMAGE_TAG=8.4.0 CONNECTOR_IMAGE=${{ needs.build-push-inference.outputs.image }}-amd64 INFERENCE_IMAGE=holisticon/bpm-ai-inference:latest-cpu OPENAI_API_KEY=${{ secrets.OPENAI_API_KEY }} poetry run pytest + run: ZEEBE_TEST_IMAGE_TAG=8.4.0 CONNECTOR_IMAGE=${{ needs.build-push.outputs.image }}-amd64 INFERENCE_IMAGE=holisticon/bpm-ai-inference:latest-cpu OPENAI_API_KEY=${{ secrets.OPENAI_API_KEY }} poetry run pytest create-push-manifest: runs-on: ubuntu-latest From 88677926f773592d8b3e87aad00f56517117f912 Mon Sep 17 00:00:00 2001 From: Bennet Date: Fri, 5 Apr 2024 00:48:34 +0200 Subject: [PATCH 2/6] Update README.md --- README.md | 21 ++++++++++----------- 1 file changed, 10 insertions(+), 11 deletions(-) diff --git a/README.md b/README.md index 793ebf5..4a71cc8 100644 --- a/README.md +++ b/README.md @@ -20,18 +20,19 @@ local OCR with tesseract and local audio transcription with Whisper. All running Example usage -### 🆕 What's New in 1.0 +### 🆕 What's New +* Anthropic Claude 3 model options * Option to use small **AI models running 100% locally on the CPU** - no API key or GPU needed! * Curated models known to work well, just select from dropdown * Or use any compatible model from [HuggingFace Hub](https://huggingface.co/models) * Multimodal input: * **Audio** (voice messages, call recordings, ...) using local or API-based transcription * **Images / Documents** (document scans, PDFs, ...) using local or API-based OCR or multimodal AI models -* Ultra slim docker image (**60mb** without local AI) +* Use files from Amazon S3 or Azure Blob Storage * Logging & Tracing support with [Langfuse](https://langfuse.com) ### 🔜 Upcoming -* higher quality local and API-based OCR +* higher quality local OCR * support for local, open-access LLMs --- @@ -101,25 +102,23 @@ mkdir ./data and launch the connector runtime with a local zeebe cluster: ```bash -docker compose --profile default --profile platform up -d +docker compose --profile platform up -d ``` For Camunda Cloud, remove the platform profile. -To use the larger **inference** image that includes dependencies to run local AI model inference for decide, extract and translate, use the inference profile instead of default: +To use the **inference** extension container that includes local AI model inference implementations for decide, extract and translate, as well as local OCR, additionally use the inference profile: ```bash docker compose --profile inference --profile platform up -d ``` -#### Available Image Tags +#### Available Images Two types of Docker images are available on [DockerHub](https://hub.docker.com/r/holisticon/bpm-ai-connectors-camunda-8): -* The lightweight (**~60mb** compressed) default image suitable for users only needing the OpenAI API (and other future API-based services) - * Use `latest` tag (multiarch) -* The more heavy-weight (~500mb) inference image that contains all dependencies to run transformer AI models (and more) **locally on the CPU**, -allowing you to use the `decide`, `extract` and `translate` connectors 100% locally without any API key needed - * Use `latest-inference` tag (multiarch) +* The main image suitable for users only needing the Anthropic/OpenAI and Azure/Amazon APIs (and other future API-based services) +* An optional inference image that contains all dependencies to run transformer AI models (and more) **locally on the CPU**, +allowing you to use the `decide`, `extract` and `translate` connectors 100% locally and perform OCR without any API key needed ## 📚 Connector Documentation From 49c4aa2c8b4b686f38972e57e8246595f191e883 Mon Sep 17 00:00:00 2001 From: Bennet Date: Fri, 5 Apr 2024 00:50:55 +0200 Subject: [PATCH 3/6] Update docker-compose.yml --- docker-compose.yml | 1 - 1 file changed, 1 deletion(-) diff --git a/docker-compose.yml b/docker-compose.yml index 9709de3..36bdb59 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -5,7 +5,6 @@ services: ##################################################################################################################### connectors: - profiles: [default] container_name: bpm-ai-connectors-camunda-8 image: holisticon/bpm-ai-connectors-camunda-8:latest build: From 46868ac3fdf37404842c101d9cdc4ebfda784a9c Mon Sep 17 00:00:00 2001 From: Bennet Date: Fri, 5 Apr 2024 00:51:36 +0200 Subject: [PATCH 4/6] Update docker-compose.yml --- docker-compose.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docker-compose.yml b/docker-compose.yml index 36bdb59..e7d0ab3 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -1,7 +1,7 @@ services: ##################################################################################################################### - # Default AI Connectors - enabled by profile `--profile default` # + # AI Connectors # ##################################################################################################################### connectors: From ebe38f5246f16b03e976459bec0be382761675da Mon Sep 17 00:00:00 2001 From: Bennet Date: Fri, 5 Apr 2024 00:53:42 +0200 Subject: [PATCH 5/6] Update wizard.sh --- wizard.sh | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/wizard.sh b/wizard.sh index abcdfac..7c21f12 100644 --- a/wizard.sh +++ b/wizard.sh @@ -78,8 +78,6 @@ inference=${inference:-n} if [ "$inference" = "y" ]; then profile_flags="$profile_flags --profile inference" -else - profile_flags="$profile_flags --profile default" fi ############################################################################################################################## @@ -215,4 +213,4 @@ done # Start docker compose with selected profile(s) ############################################################################################################################## -eval "docker compose$profile_flags up -d" \ No newline at end of file +eval "docker compose$profile_flags up -d" From fe5d2095aa027358789e5960ecad6b242d038993 Mon Sep 17 00:00:00 2001 From: Bennet Date: Fri, 5 Apr 2024 10:03:47 +0200 Subject: [PATCH 6/6] Update docker-compose.yml --- docker-compose.yml | 1 + 1 file changed, 1 insertion(+) diff --git a/docker-compose.yml b/docker-compose.yml index e7d0ab3..78740a8 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -12,6 +12,7 @@ services: depends_on: bpm-ai-inference: condition: service_started + required: false zeebe: condition: service_healthy required: false