Deploying a containerized agent built with the Google Agent Development Kit (ADK) that uses VertexAI API
This tutorial guides you through deploying a containerized agent built with the Google Agent Development Kit (ADK) to Google Kubernetes Engine (GKE). The agent uses VertexAI to access LLMs. GKE provides a managed environment for deploying, managing, and scaling your containerized applications using Google infrastructure.
Building Agents with Agent Development Kit (ADK) on GKE Autopilot cluster using Self-Hosted LLM
This tutorial demonstrates how to deploy the Llama-3.1-8B-Instruct model on Google Kubernetes Engine (GKE) and vLLM for efficient inference. Additionally, it shows how to integrate an ADK agent to interact with the model, supporting both basic chat completions and tool usage. The setup leverages a GKE Autopilot cluster to handle the computational requirements.
Building Agents with Agent Development Kit (ADK) on GKE using Ray Serve for Self-Hosted LLMs
This tutorial demonstrates how to deploy the Llama-3.1-8B-Instruct model on Google Kubernetes Engine (GKE) using Ray Serve and vLLM for efficient inference. Additionally, it shows how to integrate an ADK agent to interact with the model, supporting both basic chat completions and tool usage. The setup leverages a GKE Standard cluster with GPU-enabled nodes to handle the computational requirements.
Deploying MCP Servers on GKE: Building AI Agents with ADK and Ray-Served Models
This guide provides instructions for deploying a Ray cluster with the AI Device Kit (ADK) and a custom Model Context Protocol (MCP) server on Google Kubernetes Engine (GKE). It covers setting up the infrastructure with Terraform, containerizing and deploying the Ray Serve application, deploying a custom MCP server for real-time weather data, and finally deploying an ADK agent that utilizes these components. The guide also includes steps for verifying deployments and cleaning up resources.
Building a Multi-Agent Code Development Flow with Flowise on GKE
This tutorial will provide instructions on how to deploy and use FlowiseAI on GKE (Google Kubernetes Engine) to build and operate AI applications using a low-code/no-code approach.