Vertex AI
Build, deploy, and scale machine learning (ML) models faster, with fully managed ML tools for any use case.
Through Vertex AI Workbench, Vertex AI is natively integrated with BigQuery, Dataproc, and Spark. You can use BigQuery ML to create and execute machine learning models in BigQuery using standard SQL queries on existing business intelligence tools and spreadsheets, or you can export datasets from BigQuery directly into Vertex AI Workbench and run your models from there. Use Vertex Data Labeling to generate highly accurate labels for your data collection.
Vertex AI Agent Builder enables developers to create and deploy enterprise-grade generative AI applications. It offers both no-code and code-first approaches, allowing users to build AI agents using natural language instructions or by leveraging frameworks like LangChain and LlamaIndex.
Learn more
StackAI
StackAI is an enterprise AI automation platform to build end-to-end internal tools and processes with AI agents in a fully compliant and secure way. Designed for large organizations, it enables teams to automate complex workflows across operations, compliance, finance, IT, and support without heavy engineering.
With StackAI you can:
• Connect knowledge bases (SharePoint, Confluence, Notion, Google Drive, databases) with versioning, citations, and access controls.
• Deploy AI agents as chat assistants, advanced forms, or APIs integrated into Slack, Teams, Salesforce, HubSpot, or ServiceNow.
• Govern usage with enterprise security: SSO (Okta, Azure AD, Google), RBAC, audit logs, PII masking, data residency, and cost controls.
• Route across OpenAI, Anthropic, Google, or local LLMs with guardrails, evaluations, and testing.
• Start fast with templates for Contract Analyzer, Support Desk, RFP Response, Investment Memo Generator, and more.
Learn more
Opik
Confidently evaluate, test, and ship LLM applications with a suite of observability tools to calibrate language model outputs across your dev and production lifecycle. Log traces and spans, define and compute evaluation metrics, score LLM outputs, compare performance across app versions, and more. Record, sort, search, and understand each step your LLM app takes to generate a response. Manually annotate, view, and compare LLM responses in a user-friendly table. Log traces during development and in production. Run experiments with different prompts and evaluate against a test set. Choose and run pre-configured evaluation metrics or define your own with our convenient SDK library. Consult built-in LLM judges for complex issues like hallucination detection, factuality, and moderation. Establish reliable performance baselines with Opik's LLM unit tests, built on PyTest. Build comprehensive test suites to evaluate your entire LLM pipeline on every deployment.
Learn more