Senior AI Research Engineer

Senior AI Research Engineer

Posted Today by 1756892686

Negotiable
Outside
Remote
USA

Summary: The Senior AI Research Engineer will be instrumental in developing AI workflows and innovative agents to enhance embedded electronic system design, thereby reducing lead times for companies in the medical, consumer, and industrial sectors. The role involves leveraging generative AI advancements to automate design processes and create tailored solutions for complex challenges. The engineer will collaborate with internal teams to define infrastructure and deliver high-quality code solutions while ensuring data protection. A strong background in AI/ML development and experience with generative AI models is essential for success in this position.

Key Responsibilities:

  • Develop and deploy AI backend services to automate embedded system design processes.
  • Create new AI workflows, agents, frameworks, and tool integrations for embedded systems design tasks.
  • Define and configure LLM model infrastructure and maintain innovative performance in secure environments.
  • Create software for AI backend tools that interface with customer application infrastructure and data systems.
  • Collaborate with internal teams to understand system requirements and automate design processes.
  • Develop custom fine-tuned LLM models for specific domain knowledge in embedded systems.
  • Specify and lead high-performance computing resources for hosting LLMs and customer data.
  • Ensure timely delivery of high-quality, scalable code solutions and maintain code reliability.

Key Skills:

  • Bachelor's degree or equivalent experience in computer science, software engineering, or a related field.
  • Strong Python skills in AI/ML development (e.g., Pytorch, TensorFlow, NLP, scikit-learn).
  • Deep experience with generative AI models and tools (e.g., Llama, Claude, ChatGPT, Azure, AWS).
  • Experience with LLM model selection, fine-tuning, and AI workflows.
  • Proficiency with optimizations for accelerating LLM inference.
  • Strong collaboration skills in agile development environments.
  • Understanding of RESTful APIs, GraphQL, and databases.
  • Experience with version control, CI/CD pipelines, and containerized environments.
  • Entrepreneurial approach with a passion for innovation and customer success.

Salary (Rate): undetermined

City: undetermined

Country: USA

Working Arrangements: remote

IR35 Status: outside IR35

Seniority Level: Senior

Industry: IT

Detailed Description From Employer:

As our AI Research Engineer, you will play a pivotal role in developing AI workflows and innovative agents that accelerate embedded electronic system design, compressing lead times and helping companies create innovative medical, consumer, and industrial products faster.

In this role, you will focus on:

  • Using your knowledge of the latest advancements in generative AI to develop and deploy AI backend services that automate embedded system design (ESD)processes, enabling engineers to focus their energy on solving the toughest challenges.
  • Developing solutions to address some of the hardest problems at the pioneering of AI application.

What you get to do

  • Create new AI workflows, agents, frameworks, memory systems, and tool integrations to assist engineers with embedded systems design tasks.
  • Define and configure LLM model infrastructure; select, curate, and update models to maintain innovative performance in secure environment where protection of IP and customer data is paramount.
  • Create software for AI backend tools that interface to customer application infrastructure and data systems; develop APIs to AI models and data sources.
  • Work with internal collaborators and ESD domain authorities to understand system requirements and automate targeted design processes.
  • See opportunities for creating custom fine-tuned LLM models for specific domain knowledge in embedded systems; work with collaborators to collect, curate, and generate training data, including generating synthetic data; train and deploy custom models for ESD tasks.
  • Specify and lead high performance computing resources including on-site GPU servers for hosting LLMs, agents, and customer data.
  • Ensure timely and successful delivery of high-quality, scalable code solutions.
  • Ensure maintainability and reliability of code.

What you bring to the Team

  • Bachelor s degree or equivalent experience in computer science, software engineering, or a related field.
  • Strong Python skills in AI/ML development (eg. Pytorch, tensorflow, NLP, scikit-learn)
  • Deep experience using generative AI models (Llama, Claude, ChatGPT, Gemini, Deepseek, Mistral), tools (langchain, LlamaIndex, Haystack, Auto-GPT) and supporting infrastructure (Azure, AWS, Hugging Face, or local hosting on GPUs (NVIDIA)).
  • Experience with LLM model selection, fine-tuning, RAG, tool-use, MCP servers, AI workflows, AI agents, and prompt optimization.
  • Proficiency with vLLM, TensorRT-LLM, OpenLLM, Llama.cpp or related optimizations for accelerating LLM inference.
  • Strong ability to collaborate in agile development environments with front-end engineers, full-stack engineers, IT-support, and domain authorities in ESD across a variety of platforms (IoT, microcontrollers, FPGA, GPU).
  • Strong understanding of RESTful APIs, GraphQL, WebSockets, and databases (SQL/Postgres)
  • Experience with version control (Git), CI/CD pipelines, and containerized environments (Docker, Kubernetes).
  • Entrepreneurial approach, with a passion for innovation, customer success, and driving growth in a rapidly evolving market.

Preferred Qualifications

  • Understanding of CAD/CAE tools used in fields such as electrical engineering design and ESD.
  • Prior experience in a manufacturing, design automation, or industrial software environment.