Catalog Details
CATEGORY
deploymentCREATED BY
UPDATED AT
November 23, 2024VERSION
0.0.1
What this pattern does:
A batch workload is a process typically designed to have a start and a completion point. You should consider batch workloads on GKE if your architecture involves ingesting, processing, and outputting data instead of using raw data. Areas like machine learning, artificial intelligence, and high performance computing (HPC) feature different kinds of batch workloads, such as offline model training, batched prediction, data analytics, simulation of physical systems, and video processing. By designing containerized batch workloads, you can leverage the following GKE benefits: An open standard, broad community, and managed service. Cost efficiency from effective workload and infrastructure orchestration and specialized compute resources. Isolation and portability of containerization, allowing the use of cloud as overflow capacity while maintaining data security. Availability of burst capacity, followed by rapid scale down of GKE clusters.
Caveats and Consideration:
Ensure proper networking of components for efficient functioning
Compatibility:
Recent Discussions with "meshery" Tag
- Nov 22 | Meshery CI Maintainer: Sangram Rath
- Dec 04 | Link Meshery Integrations and Github workflow or local code
- Nov 20 | Meshery Development Meeting | Nov 20th 2024
- Nov 10 | Error in "make server" and "make ui-server"
- Nov 11 | Difference in dev Environments on port 9081 and 3000
- Nov 10 | npm run lint:fix error
- Oct 30 | Getting Meshery locally using Docker Desktop for Meshery UI contribution
- Nov 07 | Meshery + GCP Connector
- Oct 24 | Getting error when using utils.SetupContextEnv() when writing tests for relationship command
- Nov 16 | Where's the Cortex Integration of Meshmap?