InferX — Serverless GPU Inference Platform for Production Workloads

Model translategemma-27b-it-FP8-Dynamic

Namespace Model Name Type Standby GPU Standby Pageable Standby Pinned Memory GPU Count vRam (MB) CPU Memory (MB) State Revision
Trial translategemma-27b-it-FP8-Dynamic text2text Mem File File 1 32000 20.0 80000 Normal 266

Image

Prompt



Sample Rest Call

Pods

Tenant Namespace Pod Name State Required Resource Allocated Resource GPU

Logs

tenant namespace model name revision id node name create time exit info state
public Trial translategemma-27b-it-FP8-Dynamic 266 269 computeinstance-e00r2jrqynf83a8b4f 2026-03-01 17:05:27 Error("DockerContainerWaitError { error: \"\", code: 1 }") log
public Trial translategemma-27b-it-FP8-Dynamic 266 271 computeinstance-e00r2jrqynf83a8b4f 2026-03-01 17:14:33 Error("DockerContainerWaitError { error: \"\", code: 1 }") log
public Trial translategemma-27b-it-FP8-Dynamic 266 273 computeinstance-e00r2jrqynf83a8b4f 2026-03-01 17:19:31 Error("DockerContainerWaitError { error: \"\", code: 1 }") log

Snapshot History

tenant namespace model name revision nodename state detail updatetime
public Trial translategemma-27b-it-FP8-Dynamic 266 computeinstance-e00r2jrqynf83a8b4f Scheduled Scheduled 2026-03-01 17:00:26

Model Spec


Policy