CrawlJobs Logo

Research Intern - AI System Architecture Modeling and Performance

https://www.microsoft.com/ Logo

Microsoft Corporation

Location Icon

Location:
United States , Hillsboro

Category Icon

Job Type Icon

Contract Type:
Employment contract

Salary Icon

Salary:

6710.00 - 13270.00 USD / Month

Job Description:

Research Internships at Microsoft provide a dynamic environment for research careers with a network of world-class research labs led by globally-recognized scientists and engineers, who pursue innovation in a range of scientific and technical disciplines to help solve complex challenges in diverse fields, including computing, healthcare, economics, and the environment. The Azure Hardware and Systems Infrastructure organization is central to defining Microsoft's first-party Artificial Intelligence (AI) infrastructure architecture and strategy. This is a dynamic and fast-paced environment that in close partnership with sister organizations helps define System on Chip (SoC) designs, interconnect topologies, memory hierarchies, and much more, all in the context of enabling and optimizing workload optimized data flows for large-scale AI models. This organization plays a critical role in roadmap definition all the way from concept to silicon to hyperscale integration.

Job Responsibility:

  • Research Interns put inquiry and theory into practice. Alongside fellow doctoral candidates and some of the world’s best researchers, Research Interns learn, collaborate, and network for life
  • Research Interns not only advance their own careers, but they also contribute to exciting research and development strides
  • During the 12-week internship, Research Interns are paired with mentors and expected to collaborate with other Research Interns and researchers, present findings, and contribute to the vibrant life of the community
  • As a Research Intern, you will be at the forefront of hardware/software co-design and have a direct impact in answering critical questions around designing an optimized AI system and evaluating real-world impact on the Azure’s supporting hyperscale infrastructure
  • This role will evaluate opportunities to co-optimize central processing unit (CPU), graphics processing unit (GPU) and networking infrastructure for the Maia accelerator ecosystem
  • You will be expected to identify system stress points, propose novel architectural ideas, and create methodologies using a combination of workload characterization, modeling and benchmarking to evaluate their effectiveness

Requirements:

  • Accepted or currently enrolled in a PhD program in Computer Science or related STEM field
  • At least 1 year of experience with performance analysis tools and methodologies, optimization and modeling
  • Proficiency with frameworks such as PyTorch, SGLang, Dynamo, and AI accelerator programming models/compilers such as CUDA and Triton
  • Deep understanding of GPU and AI architectures including memory hierarchies, compute-communication interplay, kernel scheduling and interconnect properties
  • Familiarity with CPU/server architectures including understanding of PCIe topologies and accelerator/NIC/peripheral demand. Solid understanding of CPU involvement in dispatching, scheduling and orchestration of input data pipelines to AI accelerators
  • Hands-on experience with benchmarking, profiling, identifying perf bottlenecks and performance analysis and optimization, including trace generation, event monitoring and instrumentation
  • Familiarity with roofline performance modeling, detailed performance simulations and awareness of speed vs accuracy tradeoffs in various performance modeling methodologies
  • Ability to apply the appropriate performance analysis methodology including devising new or combinatorial approaches in evaluating complex system architecture what-if scenarios
  • Solid verbal and written communication skills

Additional Information:

Job Posted:
March 25, 2026

Employment Type:
Fulltime
Work Type:
On-site work
Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Research Intern - AI System Architecture Modeling and Performance

AI Research Lab Research Associate

We are currently seeking highly qualified interns to accelerate research towards...
Location
Location
United States , Milpitas
Salary
Salary:
43.27 - 93.15 USD / Hour
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
May 26, 2026
Flip Icon
Requirements
Requirements
  • Pursuing PhD degree (or other degree with significant research and innovation experience) in a relevant discipline (e.g. machine learning, computer science, electrical engineering, math, statistics, etc.)
  • Track record of world-class innovative contributions and ideas in machine learning
  • Experience with innovative solution development, such as developing proofs-of-concept, first-of-a-kind solutions, and/or technology transfer
  • Experience in deep learning research
  • Experience in developing deep learning software with high proficiency in data structures and algorithms
  • Strong programming skills and experience with Python, C/C++, and preferably Java
  • Software development experience in Deep Learning, GPU acceleration, and Model Optimization
  • Experience in Deep Learning and Machine Learning frameworks and models like Tensorflow, PyTorch
  • Experience in Transformer Neural Network architectures for Generative AI and natural language processing
  • Experience with Agentic AI and Generative AI workflows - desired
Job Responsibility
Job Responsibility
  • Conduct research and come up with solutions with a fast turnaround time
  • Build the software and applications for Neural Networks and Machine Learning
  • Work with system programming, Deep Learning frameworks and models, GPU acceleration, Model optimization, real-time streaming data, distributed computing, and deployment
  • Provide thought leadership and technical influence both internally and externally to HPE
  • Collaborate with HPE Labs research teams as well as external partners
  • Work in alignment with HPE's broader innovation community.
What we offer
What we offer
  • Health & Wellbeing benefits including physical, financial and emotional wellbeing support
  • Personal and professional development programs
  • Unconditional inclusion and flexibility to manage work and personal needs.
  • Fulltime
Read More
Arrow Right

Director, Digital Ecosystem Applications

This position is responsible for the Software Platforms group at the Innovation ...
Location
Location
United States , Belmont
Salary
Salary:
240000.00 - 285000.00 USD / Year
https://www.volkswagen-group.com Logo
Volkswagen AG
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 10+ years with 2+ years in a technical leadership role
  • CS, EE, M.S. Engineering (or equivalent) REQUIRED
  • M.S. Engineering (or equivalent) or PhD PREFERRED
  • Analytical and conceptual thinking – using logic and reason, creative and strategic
  • Communication skills – interpersonal, presentation and written
  • Managing interdisciplinary teams on individual projects
  • Integration – joining people, processes or systems
  • Influencing and negotiation skills
  • Problem solving
  • Resource management
Job Responsibility
Job Responsibility
  • Define the technical mission, architecture strategy, and long‑term platform vision for the In‑Vehicle Computing & Digital Ecosystem Applications team, spanning Android Automotive OS (AAOS), in‑vehicle compute platforms, Software‑Defined Vehicle (SDV) architecture, and AI‑driven cockpit intelligence
  • Provide technical leadership across the full software stack, including Android Framework, System Services, HAL layers, middleware, connectivity stacks, media/audio frameworks, HMI toolchains, and cloud‑connected AI runtimes within an SDV‑aligned architecture
  • Lead and mentor engineering teams in platform bring‑up, system integration, performance optimization, and development of AI‑agentic features, multimodal interaction models, and next‑generation speech technologies
  • Manage multi‑year budgets for platform development, AI integration, SDV‑aligned compute evolution, SoC evaluations, cloud services, and prototype programs
  • Deliver executive‑level technical reporting on architecture decisions, platform readiness, SDV integration milestones, AI progress, risks, and strategic recommendations
  • Drive strategic planning for ICC’s infotainment and cockpit portfolio, including AAOS evolution, hybrid cloud/edge AI pipelines, intelligent mobile agent technologies, and SDV‑centric software and compute roadmaps
  • Align technical roadmaps with global VW Group Innovation teams across infotainment, connectivity, AI/ML, vehicle architecture, cloud services, and SDV platform strategy, ensuring cross‑platform consistency and shared component reuse
  • Build strategic relationships with SoC vendors, Tier‑1 suppliers, cloud providers, and AI technology partners to influence cockpit compute and SDV platform evolution
  • Maintain partnerships with Silicon Valley companies specializing in AI runtimes, LLMs, speech, multimodal interaction, and automotive‑grade SDV‑compatible software frameworks
  • Collaborate with academic and research institutions on AI‑agentic systems, embedded ML, HMI, and in‑vehicle compute architectures aligned with SDV principles
What we offer
What we offer
  • Eligibility for annual performance bonus
  • Healthcare benefits
  • 401(k), with company match
  • Defined contribution retirement program
  • Tuition reimbursement
  • Company lease car program
  • Paid time off
  • Fulltime
Read More
Arrow Right

Product Marketing Manager for AI Cloud Providers and Foundation Model Builders

Product Marketing Manager for AI Cloud Providers and Foundation Model Builders –...
Location
Location
United States
Salary
Salary:
Not provided
vastdata.com Logo
VAST Data
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 8+ years of proven experience in product or solution marketing within cloud services, AI/ML platforms, large-scale data infrastructure, data storage, operating systems, or related fields, with a strong understanding of 'as-a-Service' GTM models and/or the needs of large-scale AI research and development
  • Deep understanding of infrastructure for cloud service providers and/or large-scale AI training/inference clusters, including multi-tenant architectures, service orchestration, distributed computing, virtualization, containers, and unified file/object storage solutions at petabyte to exabyte scale
  • Familiarity with modern AI/ML data pipelines, the lifecycle of Foundation Models (data ingestion, pre-processing, training, fine-tuning, inference), and analytics workloads, and how they are delivered as cloud services or built in dedicated environments
  • Strong expertise in aligning complex technical solutions like the VAST AI OS with business-driven objectives for both AI Cloud Providers (e.g., service differentiation, new revenue streams, TCO reduction) and Foundation Model Builders (e.g., faster time-to-model, research breakthroughs, efficient scaling, optimized resource utilization)
  • Exceptional communication skills, with proven ability to articulate technical concepts clearly to both technical and business audiences
  • Demonstrated success developing and executing impactful marketing strategies and campaigns
  • Highly collaborative with a proactive approach to managing cross-functional projects
  • Willingness to travel for customer engagements, industry events, and internal meetings
Job Responsibility
Job Responsibility
  • Develop and execute strategic go-to-market plans tailored to AI Cloud Providers and Foundation Model Builders for the VAST AI OS
  • Craft compelling, differentiated messaging and positioning for the VAST AI OS that resonates with stakeholders at AI Cloud Providers (product management, service architects, business development) and at organizations building Foundation Models (AI researchers, MLOps engineers, data scientists, infrastructure leads)
  • Conduct market analysis, identifying trends, threats, and opportunities in the AI cloud services, large-scale AI model development, and underlying data infrastructure landscape, relevant to the VAST AI OS
  • Translate complex technical features of the VAST AI OS into clear benefits for AI Cloud Providers (service differentiation, revenue opportunities, TCO) and for Foundation Model Builders (accelerated training, reduced data management overhead, faster iteration cycles, scalable deployment)
  • Serve as an expert resource on architectures for AI cloud services and large-scale model development, including multi-tenancy, service orchestration, distributed training, high-performance data pipelines, and how the VAST AI OS underpins these for AI-as-a-Service and Foundation Model lifecycles, emphasizing unified file and object storage, data protection, compliance, and security
  • Collaborate closely with Product Management and Engineering to influence the VAST AI OS roadmap and direction based on the unique requirements of AI Cloud Providers, Foundation Model Builders, and market insights for AI services and model development
  • Create high-impact sales tools, presentations, reference architectures, product demonstrations, webinars, and training materials for the VAST AI OS that effectively communicate technical and business advantages to and through AI Cloud Providers, and directly to organizations building Foundation Models
  • Support partner development, sales teams, and direct engagement efforts with strategic responses to AI Cloud Provider opportunities, Foundation Model initiatives leveraging the VAST AI OS, and joint RFI/RFPs
  • Engage regularly with AI Cloud Providers and key players in the Foundation Model ecosystem to capture insights, validate VAST AI OS positioning, and foster advocacy and joint marketing opportunities
  • Produce influential content for the VAST AI OS including whitepapers, case studies, solution briefs, blogs, and FAQs tailored to AI Cloud Provider and Foundation Model Builder audiences and their respective customers or users
  • Fulltime
Read More
Arrow Right

Performance Architect

In this position, you will develop AI Storage Solutions based advanced system ar...
Location
Location
United States , Milpitas
Salary
Salary:
136537.00 - 193442.00 USD / Year
sandisk.com Logo
Sandisk
Expiration Date
April 28, 2026
Flip Icon
Requirements
Requirements
  • Bachelors or Masters or PhD in Computer/Electrical Engineering with 5+ years of relevant experience in Performance Modeling, Simulation, and Analysis using SystemC
  • At least 5+ years of experience with SystemC modeling
  • Good understanding of computer/graphics architecture, ML, LLM
  • Experience of simulation using System C and TLM, behavioral modeling and performance analysis
Job Responsibility
Job Responsibility
  • Build SystemC performance models for AI Storage Solutions based products covering end-to-end from GPU/TPU/NPU/xPU, host interface, memory hierarchy, basedie controller, and AI Storage Solutions using various packaging technolgies
  • Responsible for improving the AI/ML ASIC Architecture performance through hardware & software co-optimization, post-silicon performance analysis, and influencing the strategic product roadmap
  • Workload analysis and characterization of ASIC and competitive datacenter and AI solutions to identify opportunities for performance improvement in our products
  • Collaboration with Architecture team to resolve performance issues and optimize the performance and TCO of their AI Storage Solutions based datacenter technologies
  • Experience modeling one or some components of AI/ML accelerator ASICs such as AI Storage Solutions, PCIe/UCIe/CXL, NoC, DMA, Firmware Interactions, NAND, xPU, fabrics, etc
  • Performance modeling and optimization for multi-trillion parameter LLM training/inference including Dense, Mixture of Experts (MoE) with multiple modalities (text, vision, speech)
  • Model/optimize novel parallelization strategies across tensor, pipeline, context, expert and data parallel dimensions
  • Architect memory-efficient training systems utilizing techniques like structured pruning, quantization (MX formats), continuous batching/chunked prefill, speculative decoding
  • Incorporate and extend SOTA models such as GPT-4, Reasoning models like Deepseek-R1, and multi-modal architectures
  • Collaborate with internal and external stakeholders/ML researchers to disseminate results and iterate at rapid pace
What we offer
What we offer
  • Short-Term Incentive (STI) Plan
  • Long-Term Incentive (LTI) program (restricted stock units (RSUs) or cash equivalents)
  • RSU awards for eligible new hires
  • Paid vacation time
  • Paid sick leave
  • Medical/dental/vision insurance
  • Life, accident and disability insurance
  • Tax-advantaged flexible spending and health savings accounts
  • Employee assistance program
  • Other voluntary benefit programs such as supplemental life and AD&D, legal plan, pet insurance, critical illness, accident and hospital indemnity
  • Fulltime
Read More
Arrow Right

Member of Technical Staff, Research

As a Member of Technical Staff on the Research team, you’ll push the boundaries ...
Location
Location
United States , San Mateo
Salary
Salary:
175000.00 - 240000.00 USD / Year
fireworks.ai Logo
Fireworks AI
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Research background in Artificial Intelligence, Machine Learning, Physics, or similar field
  • Experience solving analytical problems using analytic and quantitative approaches
  • Experience communicating research to audiences with different backgrounds
  • Experience coding in C/C++, Python, or other similar languages
Job Responsibility
Job Responsibility
  • Conduct foundational research to advance the capabilities, efficiency, and reliability of LLMs and multimodal systems
  • Design, implement, and evaluate novel model architectures, training methods, and optimization techniques
  • Collaborate with engineering teams to transition research prototypes into production-grade systems
  • Analyze empirical results, identify performance bottlenecks, and iterate quickly to improve model quality
  • Contribute to internal research strategy by identifying high-impact opportunities and emerging trends in AI
What we offer
What we offer
  • Meaningful equity in a fast-growing startup
  • Competitive salary
  • Comprehensive benefits package
  • Fulltime
Read More
Arrow Right

AI Engineer

Reporting to the AI & Technology Oversight Manager, the AI Engineer is responsib...
Location
Location
India , Mumbai
Salary
Salary:
Not provided
waystone.com Logo
Waystone Governance Ltd.
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Deep understanding of the distinction between Generative AI and Agentic AI, including their foundations, capabilities, and appropriate use cases
  • Strong understanding of AI, ML and LLM concepts, including prompt engineering, prompt grounding, iterative loop techniques, context windows, embeddings, RAG, agentic workflows
  • Proven ability to integrate AI capabilities both into low-code automation flows and high-code stacks, including, applications, APIs, microservices, distributed systems, and development or testing tools
  • Solid software development background with hands-on coding experience in one or more engineering ecosystem such as .NET (C#), Python, or TypeScript
  • Excellent communication skills, with the ability to translate complex AI concepts for non‑experts and to effectively influence and collaborate with stakeholders at all levels, both technical and non‑technical
  • Strong writing skills, with the ability to contribute to AI literacy and AI fluency documentation
  • Strong understanding of responsible AI principles, including governance, bias mitigation, compliance, and risk-based decision-making
  • Analytical thinking with excellent problem‑solving ability and keen attention to details
  • Ability to mentor developers and testers, and to drive innovation across engineering, QA, and architecture
  • Ability to assess AI‑enabled capabilities in third‑party SaaS platforms (e.g., Appian, Salesforce,etc) and provide guidance on responsible, effective adoption
Job Responsibility
Job Responsibility
  • Hands-on contributor to the design and development of AI-enabled solutions, capable of writing both production-quality code and rapid experimental prototypes
  • Develop and implement AI‑enabled microservices, APIs, applications, and internal tools
  • Integrate AI capabilities following secure, scalable engineering best practices
  • Design, build and validate AI‑driven solutions leveraging providers such as OpenAI and Anthropic
  • Enhance low‑code/no‑code automation platforms (e.g., Power Automate, n8n, Workato) by embedding intelligent processing and applying agentic patterns where relevant
  • Implement Model Context Protocol (MCP) servers for secure AI‑to‑system connectivity
  • Lead AI‑based document parsing and intelligent data extraction initiatives
  • Contribute to educating and enabling Enterprise Capabilities areas, including Integration and Automation, by providing guidance, training, and best practices, e.g., on effective use of n8n agents
  • Engage with business stakeholders to understand requirements, constraints, and key drivers, identifying and implementing high‑value AI opportunities across Waystone
  • Prototype AI features and iterate towards production‑ready capabilities
  • Fulltime
Read More
Arrow Right

AI Engineer

Reporting to the AI & Technology Oversight Manager, the AI Engineer is responsib...
Location
Location
United Kingdom , Leeds
Salary
Salary:
Not provided
waystone.com Logo
Waystone Governance Ltd.
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Deep understanding of the distinction between Generative AI and Agentic AI, including their foundations, capabilities, and appropriate use cases
  • Strong understanding of AI, ML and LLM concepts, including prompt engineering, prompt grounding, iterative loop techniques, context windows, embeddings, RAG, agentic workflows
  • Proven ability to integrate AI capabilities both into low-code automation flows and high-code stacks, including, applications, APIs, microservices, distributed systems, and development or testing tools
  • Solid software development background with hands-on coding experience in one or more engineering ecosystem such as .NET (C#), Python, or TypeScript
  • Excellent communication skills, with the ability to translate complex AI concepts for non‑experts and to effectively influence and collaborate with stakeholders at all levels, both technical and non‑technical
  • Strong writing skills, with the ability to contribute to AI literacy and AI fluency documentation
  • Strong understanding of responsible AI principles, including governance, bias mitigation, compliance, and risk-based decision-making
  • Analytical thinking with excellent problem‑solving ability and keen attention to details
  • Ability to mentor developers and testers, and to drive innovation across engineering, QA, and architecture
  • Ability to assess AI‑enabled capabilities in third‑party SaaS platforms (e.g., Appian, Salesforce,etc) and provide guidance on responsible, effective adoption
Job Responsibility
Job Responsibility
  • Hands-on contributor to the design and development of AI-enabled solutions, capable of writing both production-quality code and rapid experimental prototypes
  • Develop and implement AI‑enabled microservices, APIs, applications, and internal tools
  • Integrate AI capabilities following secure, scalable engineering best practices
  • Design, build and validate AI‑driven solutions leveraging providers such as OpenAI and Anthropic
  • Enhance low‑code/no‑code automation platforms (e.g., Power Automate, n8n, Workato) by embedding intelligent processing and applying agentic patterns where relevant
  • Implement Model Context Protocol (MCP) servers for secure AI‑to‑system connectivity
  • Lead AI‑based document parsing and intelligent data extraction initiatives
  • Contribute to educating and enabling Enterprise Capabilities areas, including Integration and Automation, by providing guidance, training, and best practices, e.g., on effective use of n8n agents
  • Engage with business stakeholders to understand requirements, constraints, and key drivers, identifying and implementing high‑value AI opportunities across Waystone
  • Prototype AI features and iterate towards production‑ready capabilities
  • Fulltime
Read More
Arrow Right

Research Associate

The role involves accelerating research towards new applications, core methodolo...
Location
Location
United States , Milpitas
Salary
Salary:
43.27 - 93.15 USD / Hour
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
May 26, 2026
Flip Icon
Requirements
Requirements
  • Pursuing a PhD degree (or other degree with significant research and innovation experience) in a relevant discipline (e.g. electrical engineering, computer science, machine learning, applied physics, mathematics, statistics, etc.)
  • Track record of world-class innovative contributions and ideas in AI/ML and/or related areas
  • Experience with innovative solution development, such as developing proofs-of-concept, first-of-a-kind solutions, and/or technology transfer
  • Experience with in-memory computing accelerators
  • Experience with micro architecture design for custom accelerators
  • Experience in deep learning research, algorithms, and data structures
  • Experience in design and test of custom integrated circuit IP for AI/ML applications
  • Experience with emerging analog memory devices for computing applications such as memristor, ReRAM, PCM, and others
  • Experience in system software, GPU acceleration, DL model execution and performance optimization
  • Self-motivated, proactive, with leadership qualities
Job Responsibility
Job Responsibility
  • Investigation of high-performance accelerators which combine CMOS and emerging ReRAM device technologies (or memristors) for computing applications including machine learning, neuromorphic computing, network security, finite automata, and other novel computational models
  • Design of prototype systems and/or integrated circuits
  • Invention of new architectures, circuits, and/or devices to take advantage of physical hardware systems for acceleration of target computations
  • Operation of existing hardware platforms
  • Performance evaluations with competing systems
  • Provide thought leadership and technical influence both internally and externally to HPE
  • Contribute along the full range from initial novel ideas to design, development, implementation, evaluation, and technology transfer
  • Collaborate with HPE Labs research teams as well as with external partners
  • Work in alignment with HPE's broader innovation community
What we offer
What we offer
  • Comprehensive suite of benefits supporting physical, financial and emotional wellbeing
  • Programs catered to helping employees reach career goals
  • Unconditional inclusion in work environment
  • Fulltime
Read More
Arrow Right