Field Solutions Architect, Applied AI, Google Cloud
As a Field Solutions Architect (FSA) in Applied AI, you are the "Agent Engineer" and the primary delivery arm for our customers' most critical AI initiatives. You take initial conversational prototypes and transform them into production-ready solutions, owning the end-to-end engineering life-cycle, including the transition from "Art of the Possible" to real-world business value and scalable, secure AI systems. This is a high-travel, high-impact role focused on leading technical delivery for Conversational AI pilots and establishing the first Customer User Journeys (CUJs) for our largest customers at their sites. Your role requires an understanding of software engineering, MLOps, and cloud infrastructure.Google Cloud accelerates every organization’s ability to digitally transform its business and industry. We deliver enterprise-grade solutions that leverage Google’s technology, and tools that help developers build more sustainably. Customers in more than 200 countries and territories turn to Google Cloud as their trusted partner to enable growth and solve their most critical business problems.
The US base salary range for this full-time position is $147,000-$216,000 + bonus + equity + benefits. Our salary ranges are determined by role, level, and location. Within the range, individual pay is determined by work location and additional factors, including job-related skills, experience, and relevant education or training. Your recruiter can share more about the specific salary range for your preferred location during the hiring process.
Please note that the compensation details listed in US role postings reflect the base salary only, and do not include bonus, equity, or benefits. Learn more about benefits at Google.
Responsibilities
- Lead the development of Conversational AI and CX applications, transitioning from rapid prototypes to production-grade agentic workflows (e.g., multi-agent systems, MCP servers) that drive measurable ROI.
- Architect and code conversational flows that are not just functional, but optimized for the "connective tissue" between Google’s Conversational AI products and customers’ live infrastructure, including APIs, legacy data silos, and security perimeters.
- Build high-performance evaluation (Eval) pipelines and observability frameworks to optimize agentic workloads, focusing on reasoning loops, tool selection, and reducing latency while maintaining production-grade security and networking.
- Identify repeatable field patterns and technical "friction points" in Google’s AAI stack, converting them into reusable modules or product feature requests for engineering teams.
- Co-build with customer engineering teams to instill Google-grade development best practices, ensuring long-term project success and high end-user adoption.
Minimum qualifications:
- Bachelor’s degree in Computer Science or equivalent practical experience in Software Engineering, Site Reliability Engineering (SRE), or DevOps.
- 6 years of experience in Python and experience architecting scalable AI systems on cloud platforms.
- Experience deploying conversational agents using code-based frameworks and build in real-time with customers utilizing modern generative AI tools.
- Experience in deploying resources via Terraform or similar tools to automate the setup of agents, functions, and networking.
- Experience building full-stack applications (not just scripts) that interact with enterprise IT infrastructures also in developing and driving customer projects forward in a timely manner.
Preferred qualifications:
- Master’s or PhD in AI, Computer Science, or a related technical field.
- Experience implementing multi-agent systems using frameworks (e.g., LangGraph, CrewAI, or Google’s ADK) and patterns like ReAct, self-reflection, and hierarchical delegation.
- Experience debugging Agent logic (ReAct loops, Chain of Thought) and optimizing tool selection, including tracing conversation IDs across microservices to identify and resolve failures in real-time.
- Experience connecting agents to enterprise knowledge bases and optimizing Retrieval-augmented generation (RAG) chunking to prevent hallucinations.
- Ability to troubleshoot live, high-traffic systems during critical windows.
- Ability to travel up to 50% of the time.