CBorg API vs Cloud Project

Note

Got a question? Email us at scienceit@lbl.gov or join the CBorg Users Chatroom

There are two primary ways to access LLM models at Berkeley Lab: through the CBorg API Proxy or through a self-managed Cloud Project (GCP or AWS). This page compares the two approaches to help you decide which is the best fit for your use case.


CBorg API Proxy

The CBorg API Proxy is a managed service operated by the IT Division. It provides a single OpenAI-compatible API endpoint that routes requests to a variety of commercial and on-premises models.

Key Features:

  • More models available — includes OpenAI, xAI, Google Gemini, and Anthropic Claude models through a single endpoint
  • Managed model list — we keep the list of available models up to date as new versions are released
  • Budget enforcement — we manage budget controls so you don’t accidentally exceed your spending limit
  • $50/month in free usage per user — every user gets a free monthly allocation
  • On-prem zero-cost models — the following models are hosted on-premises and have no per-token cost:
    • gpt-oss-120b
    • gemini-4
    • nomic-embed-text
    • nomic-embed-code

Considerations:

  • May go offline briefly for maintenance
  • Acceptable for research purposes but not intended to support “critical infrastructure”

Cloud Project (GCP / AWS)

A Cloud Project gives you a dedicated GCP or AWS account attached to your PID, where you manage your own model access and configuration.

Key Features:

  • Lower inference cost — cost of LLM inference may be around 10% lower than the “retail” cost charged by CBorg
  • High availability — >99.99% uptime; cloud services will almost never go offline
  • Data isolation — your prompts do not flow through the CBorg system
  • Dedicated inference bandwidth — if you need to run many prompts in parallel at a high sustained rate
  • $25/month in free usage per project

Considerations:

  • Self-managed — you are responsible for:
    • Setting up and monitoring budget alerts and cost controls
    • Setting up API keys / service accounts to access models
    • Deploying new models when they come out
    • Generally being responsible for configuration and security
  • No access to OpenAI models — Gemini, Claude, AWS Nova, Cohere and others are available, but OpenAI models are not

Summary Comparison

FeatureCBorg API ProxyCloud Project
Available ModelsOpenAI, xAI, Gemini, Claude, on-premGemini, Claude, AWS Nova, Cohere, others (no OpenAI)
Free Usage$50/month per user$25/month per project
Model ManagementManaged by IT DivisionSelf-managed
Budget ControlsManaged by IT DivisionSelf-managed
On-Prem Models✅ Zero cost❌ Not available
UptimeMay have brief maintenance windows>99.99%
Data IsolationPrompts routed through CBorgFull isolation
Inference CostStandard retail pricing~10% lower
Parallel ThroughputShared bandwidthDedicated bandwidth
Use CaseResearch, prototyping, general useCritical infrastructure, high-throughput, sensitive data

Which should I choose?

  • Choose CBorg API if you want a simple, managed experience with access to the widest range of models including OpenAI, free on-prem models, and automatic budget controls.
  • Choose a Cloud Project if you need high availability for critical infrastructure, dedicated inference bandwidth, data isolation, lower per-token costs at scale, or processing of sensitive data (ECI, CUI) with approval from IT Policy.

To get started with a Cloud Project, email scienceit@lbl.gov.