CrawlJobs Logo

Senior Solution Architect AI & HPC

https://www.hpe.com/ Logo

Hewlett Packard Enterprise

Location Icon

Location:

Category Icon

Job Type Icon

Contract Type:
Not provided

Salary Icon

Salary:

Not provided

Job Description:

AI is a high-growth market for HPE, and we believe we are uniquely suited to bring our end-to-end technologies spanning compute technologies from Edge to Cloud, purpose-built AI software, interconnect, and services, to support our customers' AI journey. This is an incredibly exciting time to join a dynamic and rapidly growing team where you can play a key role in understanding the needs of our customers and enabling them to deliver their mission critical AI initiatives.

Job Responsibility:

  • Collaborate with sales teams to understand customer requirements and develop tailored solutions for their AI infrastructure needs
  • Engage in pre-sales activities, including technical presentations, demonstrations, and proof-of-concepts
  • Act as a trusted advisor to customers, addressing their questions, concerns, and technical challenges effectively
  • Stay up-to-date with the latest advancements in AI technologies, cloud architectures, and infrastructure trends
  • Lead Proof-of-Concepts (PoC) for HPE customers expanding into Deep Learning or Machine Learning use cases
  • Architect reusable end-to-end AI solutions for HPE customers and prospects
  • Lead technical discussions with customers and partners to propose HPE and partner Integrated solutions
  • Identify solutions, define action plans, and help coordinate and deliver optimal solutions and enhancements
  • Recommend configurations and settings for different types of hardware and interconnect fabrics
  • Assist in any product or technical issue towards an initial sale or renewal of a customer
  • Maintains excellent communication with customer, with key focus on IT managers, administrators, and specialists
  • Work in collaboration with other Solution Engineers / Solution Architects to integrate the full solution pipeline
  • Assesses the impact of new technologies on the company's technical solution portfolio
  • Actively monitors competitive activities and their client advocates within accounts to anticipate and plans for competitive threats

Requirements:

  • Bachelor's or Master's degree in Engineering, Computer Science, or similar quantitative focus preferred
  • Ability to quickly prototype functionality into scripts for demos, integrations, troubleshooting, etc.
  • Expertise in cloud architectures, specifically with public cloud platforms such as AWS, Azure, or Google Cloud
  • Strong understanding of AI technologies, including machine learning, deep learning, and neural networks
  • Experience participating in solution configurations and the creation of PoCs to meet customer requirements
  • Solid knowledge of infrastructure components, including servers, storage, networking, and virtualization
  • Experience with high-performance computing (HPC) and GPU-accelerated systems is advantageous
  • Demonstrates expert technical skills in assigned area of specialization
  • Expert knowledge of the company offerings, strategic initiatives, current trends, competitor products and strategies within area of responsibility
  • Expert level written and verbal communication skills and mastery over English and local language
  • Demonstrates expert consultative selling techniques, including active listening, framing, white boarding, storytelling etc.
  • Broad knowledge of partner´s and non-standard partner´s offerings and how/when to leverage them for deals within area of specialization
  • Strong project management knowledge or experience with excellent analytical and problem solving skills
  • Understand the correlation of model, parameter density, data density - with resulting compute and storage requirements
  • Understand the correlation between CPU, GPUs, memory for different types of models
  • Ability to assess when is on-prem better than Cloud, or not
  • Understanding of how infrastructure scales as the model scales
  • Strong communicator, presenter and technologist evangelist
  • Curiosity/interest in continuous learning to stay at the forefront of challenges which can be addressed through AI

Nice to have:

  • Certification in cloud services is highly desirable
  • Familiarity with popular AI frameworks and libraries is a plus
What we offer:
  • Health & Wellbeing benefits
  • Personal & Professional Development programs
  • Unconditional Inclusion environment
  • Comprehensive suite of benefits that supports physical, financial and emotional wellbeing

Additional Information:

Job Posted:
September 25, 2025

Employment Type:
Fulltime
Work Type:
Hybrid work
Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Senior Solution Architect AI & HPC

Senior AI Presales Consultant

We are seeking a high-impact, strategic AI Presales Consultant to join our elite...
Location
Location
India , Mumbai
Salary
Salary:
Not provided
eviden.com Logo
Eviden
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 7+ years in a customer-facing technical role (e.g., Presales, Solutions Architecture, AI Specialist, or Technical Consulting), with a proven track record of designing large-scale AI, ML, or HPC solutions
  • Deep, hands-on understanding of LLM architectures. Must be able to architect, explain, and build PoCs for RAG pipelines, including vector databases (e.g., Milvus, Pinecone, Chroma), embedding models, and data ingestion strategies
  • Direct experience in sizing AI infrastructure. Must be able to perform "napkin math" and detailed calculations for GPU, CPU, memory, and network requirements
  • Must be able to fluently discuss performance metrics (tokens/second, latency, throughput, TFLOPS) and their relationship to hardware choice (e.g., NVIDIA H100 vs. A100, memory bandwidth, interconnects like NVLink/InfiniBand)
  • Expertise in the AI software stack. Strong understanding of MLOps principles (Kubeflow, MLflow), Kubernetes (K8s) for AI workloads, and model serving platforms (NVIDIA Triton, KServe, or similar)
  • Strong, current knowledge of the AI model landscape (e.g., Llama family, Mistral, GPT-family, foundation models). Ability to discuss fine-tuning techniques, quantization, and pruning
  • Exceptional communication, whiteboarding, and presentation skills. Ability to translate executive-level business needs into detailed technical architecture and build a compelling C-level value proposition
  • Bachelor's or Master's degree in Computer Science, AI, Data Science, or a related engineering field
Job Responsibility
Job Responsibility
  • Strategic Client Advisory: Lead executive-level "Art of the Possible" workshops and technical discovery sessions to understand a client's business goals, data readiness, and AI maturity
  • Full-Stack Solution Architecture: Design holistic, end-to-end AI solutions that synergize our supercomputing hardware, AI software platform, and MLOps capabilities to meet specific client needs
  • Generative AI & LLM Expertise: Act as the subject matter expert on Generative AI. Architect and evangelize scalable data ingestion and preparation pipelines, specializing in Retrieval-Augmented Generation (RAG) frameworks
  • Infrastructure Sizing & Performance Modelling: Analyse customer workloads (data volume, model complexity, training frequency, inference throughput) to accurately size the required platform infrastructure, including Kubernetes clusters, data storage, and software licenses. This includes calculating compute, storage, and network requirements based on key performance metrics like model parameters, token performance (tokens/sec), desired latency, and concurrent user load
  • Model & Software Consultation: Advise clients on AI model selection, comparing the trade-offs of open-source vs. proprietary LLMs, fine-tuning vs. foundation models, and model quantization
  • Position and demonstrate our proprietary AI software platform, MLOps tools, and libraries, integrating them into the client's ecosystem
  • Inference Optimization: Design and architect robust, low-latency, and high-throughput inference solutions for complex AI models, including large-scale LLM serving
  • User Experience (UX) Advocacy: Collaborate with client teams to define the end-user experience, ensuring the solution delivers tangible business value and a seamless interface for data scientists, analysts, and application users
  • Sales Cycle Enablement: Own the technical narrative throughout the sales cycle. Build and deliver compelling presentations, custom demonstrations, and Proofs of Concept (PoCs). Lead the technical response to complex RFIs/RFPs
  • Fulltime
Read More
Arrow Right

Senior AI Factory Solution Product Manager

Senior AI Factory Solution Product Manager
Location
Location
United States , Spring
Salary
Salary:
126500.00 - 276500.00 USD / Year
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
April 27, 2026
Flip Icon
Requirements
Requirements
  • Bachelor's degree in Computer Science, Engineering, Business, or a related field
  • MBA or advanced degree preferred
  • 8+ years of product management experience
  • At least 5 years focused on AI/ML products or solutions
  • Demonstrated ability to build large-scale AI solutions that bring together hardware, software and services into a cohesive offering
  • Strong understanding of AI technologies, including AI/ML lifecycle (training, tuning, inferencing), large language models, computer vision, and cloud-based AI platforms (e.g., AWS SageMaker, Microsoft AzureML, Google AI)
  • Proven track record of launching successful AI products, with experience in agile methodologies and tools like Jira
  • Background in High Performance Computing (HPC) and experience blending it with AI workloads will be an advantage
  • Excellent analytical skills, with proficiency in data analysis and market testing
  • Outstanding communication and stakeholder management abilities, capable of presenting to technical and non-technical audiences up to the senior executive/SVP levels
Job Responsibility
Job Responsibility
  • Define and drive the overall AI factory at-scale and sovereign solution vision, roadmap, and features, while closely aligning with customer needs and HPE strategic goals
  • Define and drive the key software components necessary for the solution, which may be a mix of HPE developed, commercial and community IP
  • Conduct market research, competitive analysis, and customer interviews to identify AI factory opportunities and validate solution ideas and software features in a quick turn manner
  • Collaborate with engineers, product managers and presales architects to translate requirements into technical specifications and prototypes. Use this data to inform the build/partner/buy decision on key SW
  • Oversee the software integration and end-to-end solution lifecycle, from feature ideation and MVP development to launch, iteration, and scaling
  • Monitor solution performance using KPIs like full-stack wins, product mix, customer satisfaction, and iterate offering based on data insights
  • Work with legal, finance, pricing and supply chain to setup and manage resale contracts for commercial SW
  • Partner with sales and marketing to develop go-to-market strategies, pricing models, support strategies and customer enablement materials
  • Ensure solution complies with ethical AI standards while ensuring highest level of data privacy and sovereignty (e.g., GDPR, CCPA)
  • Stay abreast of AI trends, such as generative models, agentic AI, and industry applications
What we offer
What we offer
  • Health & Wellbeing
  • Personal & Professional Development
  • Unconditional Inclusion
  • Fulltime
Read More
Arrow Right
New

Silicon Photonics Design Engineer

We are seeking a seasoned technical contributor that drives end-to-end delivery ...
Location
Location
United States , New Jersey
Salary
Salary:
171200.00 - 256800.00 USD / Year
amd.com Logo
AMD
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Masters or PhD degree in Applied Physics/Electrical Engineering/or related discipline
  • Experience developing PICs that have reached HVM, from concept to field failure analysis
  • Subject matter expertise in silicon photonics foundry technology, device design, and circuit layout, including FDTD/MODE/other related design tools
  • Proficient in software architecture and development for photonic design automation, test automation, and data analysis
  • Python and Cadence development experience is preferred
  • Experience in photonic wafer-, die-, and package-level testing and related test instrumentation
  • Excellent communication, management, and presentation skills
  • Adept at collaboration among top-thinkers and senior architects with strong interpersonal skills to work across teams in different geographies
Job Responsibility
Job Responsibility
  • Development of OE systems in advanced Si photonics process technologies for DWDM and CWDM applications in next generation AI and HPC applications
  • Innovation, product definition and specification of PIC architectures, TX chains, RX chains, components and test methodologies to meet demanding next generation Optical IO applications
  • Work with systems teams to devise, maintain, simulate and optimize Optical IO link budgets reflecting key product requirements, PIC architectures, channel properties, active device component and passive component limits as well as foundry process technology capabilities
  • Collaborate on design-for-test best practices to ensure that photonic devices will be ready for high-volume manufacturing
  • Participate in design reviews as a subject matter expert in silicon photonic component design
  • Work cross functionally with EIC, packaging, foundry technology, test development, and product teams to architect appropriate design and test automation solutions
Read More
Arrow Right

Senior Principal HPC/AI Architect

As a Senior Principal HPC/AI Architect at NTT DATA, you will lead the design of ...
Location
Location
Spain , Madrid
Salary
Salary:
Not provided
nttdata.com Logo
NTT DATA
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 8+ years in HPC/AI infrastructure design
  • 5+ years working with GPU-accelerated systems
  • Proven experience with large-scale GPU deployments (1000+ GPUs)
  • Successful track record in technical bid support and customer engagement
  • Technical Competencies: GPU Architectures: NVIDIA (H100, H200, B100, B200), AMD (MI300X), Intel (Gaudi2/3)
  • Interconnects: InfiniBand (HDR/NDR/XDR), NVLink, RoCE, Infinity Fabric
  • Storage Systems: Lustre, GPFS, BeeGFS, NVMe-oF, S3-compatible object storage
  • Container Platforms: Kubernetes, Docker, Singularity/Apptainer
  • Performance Tools: NVIDIA Nsight, ROCm, Intel VTune
Job Responsibility
Job Responsibility
  • AI Factory Architecture & Design (35%): Design GPU cluster architectures for AI and HPC workloads
  • Define node configurations for diverse workload types
  • Specify and validate performance metrics
  • Architect multi-tier interconnect networks
  • Develop topology designs and calculate bandwidth/latency targets
  • Model performance for customer workloads
  • Pre-Sales Technical Leadership (30%): Lead technical discussions with customer architects
  • Conduct workload sizing and architectural presentations
  • Develop technical content for proposals
  • Analyze competitor solutions
What we offer
What we offer
  • Opportunity to work on cutting-edge AI infrastructure projects
  • Collaborative and innovative work environment
  • Access to advanced lab infrastructure and vendor technologies
  • Career development through technical leadership and innovation
  • Fulltime
Read More
Arrow Right

Senior Manager, AI Infrastructure and Operations

The Sr. Manager/Staff Engineer, AI Infrastructure & MLOps Engineering is a senio...
Location
Location
Japan , Tokyo
Salary
Salary:
Not provided
pfizer.de Logo
Pfizer
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 8+ years of hands-on software engineering experience in cloud infrastructure, DevOps, and MLOps
  • Deep expertise in Python, Kubernetes, Terraform, Helm, and CI/CD pipeline development
  • Proven experience architecting and operating containerized solutions on AWS, GCP, and Azure
  • Strong knowledge of Infrastructure-as-Code, distributed systems, and production system reliability
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or related field
Job Responsibility
Job Responsibility
  • Design, implement, and own large-scale cloud-based HPC and MLOps platforms supporting AI model training, genomic sequencing, and precision medicine
  • Architect multi-environment clusters (AWS, GCP, Azure), enabling GPU/FPGA workloads and advanced observability
  • Lead the development of developer and cloud platforms, including internal engineering accelerators and reusable toolsets
  • Design, implement, and manage unified platform catalogs using Backstage, enhancing developer experience and application metadata management
  • Develop custom plugins and APIs for Backstage to support internal engineering workflows and documentation
  • Build and maintain Python-based automation frameworks, CI/CD pipelines, and Infrastructure-as-Code (Terraform, Helm, Pulumi, AWS CDK)
  • Operationalize containerized solutions using Docker and Kubernetes, integrating MLflow, Kubeflow, and other orchestration platforms
  • Implement robust automation for provisioning, configuring, and managing cloud resources across multiple environments
  • Lead the implementation of Service Level Indicators (SLIs), Service Level Objectives (SLOs), and advanced observability (Prometheus, Grafana, PagerDuty)
  • Develop and maintain APIs and services for model management, feature stores, and inference pipelines
  • Fulltime
Read More
Arrow Right

Senior Product Manager

We are hiring a foundational Product Manager to work directly with the CTO to de...
Location
Location
Israel , Ramat Gan
Salary
Salary:
Not provided
SQream
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5+ years of experience as a Product Manager or Solutions Architect in infrastructure, HPC, data systems, GPU/AI pipelines, or distributed systems
  • Strong outbound / customer-facing skills: presenting to CTOs, architects, OEM teams, GSIs, and technical buyers
  • Ability to operate at kernel-level conceptual depth and translate physics into product strategy
  • Exceptional communication skills - written and verbal - with the ability to simplify complex GPU and dataflow concepts
  • Demonstrated ability to drive roadmap execution with engineering while also leading external discovery and evangelism
  • Comfort owning both internal product discipline and external technical influence
Job Responsibility
Job Responsibility
  • Product Ownership (Internal): Work directly with the R&D to shape the GPU-native roadmap for ingestion, vectorization, transformation, curation, and continuous production flow
  • Define precise specifications, APIs, pipeline behavior, and physics-aligned constraints
  • Ensure product features adhere to SCAILIUM’s rigid boundaries: No orchestration. No system of record. No serving. No dashboards
  • Enforce documentation rigor. Documentation is code
  • Technical Outbound Leadership (External): Serve as a public-facing authority on GPU starvation, impedance incompatibility, and the AI Production Layer
  • Lead technical sessions with Partners, OEMs (Dell, Supermicro, HPE), GSIs (Accenture, Deloitte), and strategic enterprise customers
  • Conduct in-depth customer pipeline analyses to identify physical constraints and translate them into SCAILIUM features or patterns
  • Present SCAILIUM’s architecture in a clear, authoritative, physics-grounded manner
  • Support sales, partnerships, and field engineering by communicating the “why” behind every product decision
  • Build artifacts that shape the category: reference architectures, workload blueprints, TCO models, and silicon saturation narratives
Read More
Arrow Right

Senior Research Software Development Engineer, MSR AI for Science

We are on the cusp of a new frontier in which machine learning and artificial in...
Location
Location
Netherlands , Schiphol
Salary
Salary:
Not provided
https://www.microsoft.com/ Logo
Microsoft Corporation
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Master's degree or equivalent work experience in Computer Science, Physics, Engineering, Chemistry, Mathematics or a related field
  • Strong familiarity with Linux and the open-source ecosystem
  • Proficient working with large datasets in a cloud or HPC environment
  • Proficient in building and optimizing distributed systems and large-data applications, including those using tensor accelerators or GPUs
  • Strong analytical, problem-solving, and communication skills
  • Passionate about pushing the boundaries of science
Job Responsibility
Job Responsibility
  • Architect, design, and implement scalable and robust solutions for machine learning and scientific research involving large volumes of heterogeneous data
  • Build and optimize distributed data processing and model building pipelines
  • Develop and maintain tools and technologies for building, training, optimizing, scaling machine learning solutions
  • Collaborate with cross-functional teams, including scientists, researchers, and software engineers
  • Document and share best practices across the organization
  • Maintain the highest standards in code quality and software design
  • Fulltime
Read More
Arrow Right

Senior Software Development Engineer

We are seeking an experienced and highly technical SMTS Software Development Eng...
Location
Location
United Kingdom
Salary
Salary:
Not provided
amd.com Logo
AMD
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor’s or Master’s degree in Computer Science, Computer Engineering, or related technical field
  • 8+ years of software engineering experience in systems software, runtime libraries, GPU programming, or compiler/runtime interfaces
  • Strong proficiency in modern C++ (C++14/C++17 or newer), templates, memory models, and low‑level systems programming
  • Deep understanding of at least one GPU computing model (HIP, CUDA, SYCL, OpenCL, OpenMP offload)
  • Hands‑on experience with runtime systems, driver interfaces, or high‑performance compute libraries
  • Strong debugging skills using tools such as gdb, sanitizers, profilers, and GPU debugging tools
  • Solid understanding of parallel programming concepts—memory hierarchy, synchronization, concurrency, thread scheduling
Job Responsibility
Job Responsibility
  • Architect, implement, and optimize features in the HIP runtime, including memory management, kernel dispatch, device abstraction, multi‑GPU coordination, and synchronization primitives
  • Contribute to the evolution of the HIP programming model and interoperability with ROCr, HSA runtime, and compiler toolchains
  • Ensure functional correctness, performance, and scalability of runtime APIs across different GPU generations
  • Conduct root‑cause analysis and systems‑level debugging across the runtime, driver, compiler, and hardware layers
  • Profile GPU applications and internal runtime components to identify bottlenecks and design performance improvements
  • Optimize HIP runtime behavior for large-scale AI, HPC, and cloud workloads
  • Work closely with compiler teams (LLVM/Clang), driver teams, GPU architecture, and systems engineers to deliver end‑to‑end GPU software solutions
  • Contribute to API specifications and collaborate with upstream open-source communities where appropriate
  • Define and drive technical strategy for correctness, reliability, and conformance of the HIP runtime
  • Support enhancements in automated testing, CI, and stress/failure scenarios in the HIP test suite
Read More
Arrow Right