{"id":74918,"date":"2026-04-16T03:40:51","date_gmt":"2026-04-16T03:40:51","guid":{"rendered":"https:\/\/www.devopsschool.com\/blog\/senior-robotics-research-scientist-role-blueprint-responsibilities-skills-kpis-and-career-path\/"},"modified":"2026-04-16T03:40:51","modified_gmt":"2026-04-16T03:40:51","slug":"senior-robotics-research-scientist-role-blueprint-responsibilities-skills-kpis-and-career-path","status":"publish","type":"post","link":"https:\/\/www.devopsschool.com\/blog\/senior-robotics-research-scientist-role-blueprint-responsibilities-skills-kpis-and-career-path\/","title":{"rendered":"Senior Robotics Research Scientist: Role Blueprint, Responsibilities, Skills, KPIs, and Career Path"},"content":{"rendered":"\n<h2 class=\"wp-block-heading\">1) Role Summary<\/h2>\n\n\n\n<p>The <strong>Senior Robotics Research Scientist<\/strong> advances the company\u2019s robotics intelligence capabilities by inventing, validating, and transferring novel algorithms and learning-based methods into usable software components for real-world or simulated robots. The role blends deep research rigor (hypothesis-driven experimentation, publication-quality evaluation) with engineering pragmatism (reproducible code, measurable performance, integration-ready deliverables).<\/p>\n\n\n\n<p>In a software or IT organization, this role exists to build <strong>robotics AI as productizable software<\/strong>\u2014enabling robotics features within platforms (e.g., autonomy stacks, perception services, simulation tooling, fleet intelligence) and improving differentiation through proprietary algorithms and data advantages. Business value is created through higher autonomy performance, reduced development and deployment cost, faster iteration through simulation and data tooling, and IP creation (patents, defensible approaches, unique datasets).<\/p>\n\n\n\n<p>This is an <strong>Emerging<\/strong> role: the core responsibilities are real today, while the expected scope is expanding rapidly due to foundation models, synthetic data, scalable simulation, and automated evaluation pipelines.<\/p>\n\n\n\n<p><strong>Typical interaction teams\/functions<\/strong>\n&#8211; AI &amp; ML Engineering (applied ML, MLOps, model optimization)\n&#8211; Robotics Software Engineering (ROS2, runtime, integration)\n&#8211; Simulation\/Infrastructure (GPU clusters, physics sims, digital twins)\n&#8211; Product Management (robotics features, roadmap, customer needs)\n&#8211; QA\/Safety\/Validation (test strategy, safety cases where relevant)\n&#8211; Data Engineering (data pipelines, labeling, governance)\n&#8211; Security\/Privacy\/Compliance (data handling, vendor risk, model governance)\n&#8211; Customer\/Field Engineering (deployment feedback, telemetry, failure analysis)<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">2) Role Mission<\/h2>\n\n\n\n<p><strong>Core mission<\/strong><br\/>\nCreate and mature robotics intelligence\u2014perception, planning, control, and learning systems\u2014by researching and validating methods that measurably improve autonomy, robustness, safety, and scalability, then translating those methods into integration-ready software artifacts and evaluation frameworks.<\/p>\n\n\n\n<p><strong>Strategic importance to the company<\/strong>\n&#8211; Builds differentiated IP and technical advantage in robotics AI\n&#8211; Enables new robotics product capabilities (e.g., navigation in complex environments, manipulation, human-robot interaction)\n&#8211; Reduces time-to-iterate via simulation, synthetic data, and automated evaluation\n&#8211; Improves reliability and safety of autonomy behaviors, directly impacting customer trust and deployment viability\n&#8211; Establishes credibility through scientific rigor, publications, and thought leadership (where aligned to company strategy)<\/p>\n\n\n\n<p><strong>Primary business outcomes expected<\/strong>\n&#8211; Demonstrable performance improvements on key autonomy benchmarks (success rate, safety events, generalization)\n&#8211; Reduction in failure modes through better modeling, training, and evaluation\n&#8211; Research-to-product transfer: prototypes become maintainable modules, services, or SDK components\n&#8211; Scalable experimentation platform: reproducibility, dataset\/version governance, and measurable iteration velocity\n&#8211; Tangible IP output: patents, defensible datasets, or proprietary evaluation methodologies<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">3) Core Responsibilities<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Strategic responsibilities<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Define research agenda aligned to product strategy<\/strong>: Identify high-leverage robotics AI problems (e.g., sim2real robustness, manipulation generalization, uncertainty estimation) and translate them into a prioritized research roadmap.<\/li>\n<li><strong>Set technical direction for emerging robotics methods<\/strong>: Evaluate and recommend approaches (RL, imitation learning, world models, diffusion policies, vision-language-action models) based on evidence, risk, and feasibility.<\/li>\n<li><strong>Own benchmark strategy and success criteria<\/strong>: Establish objective metrics, test suites, and acceptance thresholds that align research outcomes with product outcomes.<\/li>\n<li><strong>Drive IP strategy with leadership<\/strong>: Contribute to patentable ideas, invention disclosures, and technical defensibility through unique datasets, architectures, or evaluation methods.<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">Operational responsibilities<\/h3>\n\n\n\n<ol class=\"wp-block-list\" start=\"5\">\n<li><strong>Plan and execute experiments end-to-end<\/strong>: Form hypotheses, design experiments, run ablations, manage compute budgets, and produce decision-ready findings.<\/li>\n<li><strong>Build reproducible research pipelines<\/strong>: Implement experiment tracking, dataset versioning, and reproducible training\/evaluation workflows to reduce \u201cone-off\u201d results.<\/li>\n<li><strong>Curate and improve data sources<\/strong>: Partner with data engineering to refine data collection, labeling, augmentation, synthetic data generation, and governance.<\/li>\n<li><strong>Operationalize model evaluation<\/strong>: Create automated regression testing across scenarios, distributions, and edge cases (including long-tail failures).<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">Technical responsibilities<\/h3>\n\n\n\n<ol class=\"wp-block-list\" start=\"9\">\n<li><strong>Develop and validate ML models for robotics<\/strong>: Train and evaluate perception models (3D vision, tracking), policy models (control, action prediction), and representation learning components.<\/li>\n<li><strong>Advance sim2real and domain adaptation<\/strong>: Design domain randomization, system identification strategies, and fine-tuning methods that improve transfer to physical environments.<\/li>\n<li><strong>Integrate learning with classical robotics<\/strong>: Combine ML with planners, controllers, state estimators, and safety layers for robust performance.<\/li>\n<li><strong>Optimize models for deployment constraints<\/strong>: Address latency, memory, power, and runtime reliability (quantization, distillation, inference acceleration).<\/li>\n<li><strong>Build or enhance simulation assets<\/strong>: Collaborate on physics simulation fidelity, sensor models, scene generation, and scenario libraries.<\/li>\n<li><strong>Perform failure analysis and debugging<\/strong>: Use logs, telemetry, and dataset introspection to identify root causes and propose mitigations.<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">Cross-functional or stakeholder responsibilities<\/h3>\n\n\n\n<ol class=\"wp-block-list\" start=\"15\">\n<li><strong>Partner with engineering for production transfer<\/strong>: Translate prototypes into maintainable components with clear APIs, tests, and documentation.<\/li>\n<li><strong>Communicate results to technical and non-technical audiences<\/strong>: Present findings, tradeoffs, and recommendations to product, leadership, and partner teams.<\/li>\n<li><strong>Support customer\/field learning loops<\/strong>: Use deployment feedback (where applicable) to prioritize research fixes and robustness improvements.<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">Governance, compliance, or quality responsibilities<\/h3>\n\n\n\n<ol class=\"wp-block-list\" start=\"18\">\n<li><strong>Ensure research quality and integrity<\/strong>: Maintain rigorous baselines, prevent data leakage, document methodology, and ensure reproducibility.<\/li>\n<li><strong>Contribute to model governance<\/strong>: Align with responsible AI practices, data privacy constraints, and security requirements for datasets and model artifacts.<\/li>\n<li><strong>Safety and validation alignment (context-specific)<\/strong>: Where robotics is safety-relevant, contribute to validation strategy, hazard analysis inputs, and evidence generation for safety cases.<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">Leadership responsibilities (Senior IC expectations)<\/h3>\n\n\n\n<ol class=\"wp-block-list\" start=\"21\">\n<li><strong>Technical mentorship<\/strong>: Mentor junior scientists\/engineers on experimentation practices, rigor, and robotics-specific ML techniques.<\/li>\n<li><strong>Lead cross-functional technical initiatives<\/strong>: Drive a multi-team effort (e.g., autonomy benchmark overhaul, sim2real program) with clear milestones and ownership boundaries.<\/li>\n<li><strong>Raise the technical bar<\/strong>: Define best practices for evaluation, experiment hygiene, and research-to-product handoff across the Robotics\/AI function.<\/li>\n<\/ol>\n\n\n\n<h2 class=\"wp-block-heading\">4) Day-to-Day Activities<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Daily activities<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Review experiment results, training curves, and evaluation dashboards; identify next iterations.<\/li>\n<li>Read and synthesize relevant papers, open-source implementations, and competitor\/industry signals.<\/li>\n<li>Implement model improvements or evaluation harness updates (often Python + PyTorch\/JAX; occasional C++\/CUDA).<\/li>\n<li>Debug failure cases using logs, visualization tooling, and targeted scenario replays in simulation.<\/li>\n<li>Quick alignment with engineering counterparts on integration constraints (APIs, latency budgets, sensor interfaces).<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Weekly activities<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Plan and queue larger experiments (hyperparameter sweeps, ablations, scenario expansions) and manage compute priorities.<\/li>\n<li>Attend robotics\/AI research review meeting: share results, risks, and \u201cstop\/continue\u201d decisions.<\/li>\n<li>Pair with data engineers or labeling ops to refine dataset slices and sampling strategies.<\/li>\n<li>Collaborate with simulation team on scenario generation, domain randomization settings, or sensor model changes.<\/li>\n<li>Provide mentorship\/code reviews to maintain research code quality and reproducibility.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Monthly or quarterly activities<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Refresh the research roadmap and align with product and engineering roadmaps (capability milestones, integration windows).<\/li>\n<li>Run benchmark recalibration: update baseline models, expand test suites, validate metric stability.<\/li>\n<li>Produce decision memos: technical tradeoffs, ROI estimates, and recommendations for investment.<\/li>\n<li>Contribute to patent reviews, invention disclosures, or publication submissions (as company policy allows).<\/li>\n<li>Participate in quarterly planning: define OKRs, compute budgets, and staffing needs.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Recurring meetings or rituals<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Weekly research standup (progress, blockers, next experiments)<\/li>\n<li>Biweekly cross-functional autonomy review (research + engineering + product)<\/li>\n<li>Monthly benchmark council (metrics, scenario library, regression thresholds)<\/li>\n<li>Quarterly strategy\/OKR planning with Head\/Director of Robotics or AI<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Incident, escalation, or emergency work (context-specific)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Triage critical regressions discovered in pre-release evaluation (e.g., sudden drop in navigation success rate).<\/li>\n<li>Support root-cause analysis after field incidents by producing targeted evaluation runs and hypotheses for engineering fixes.<\/li>\n<li>Urgent mitigation design (e.g., constraint layers, model fallback behavior) when a deployment risk is identified.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">5) Key Deliverables<\/h2>\n\n\n\n<p><strong>Research and technical deliverables<\/strong>\n&#8211; Research roadmap (quarterly\/biannual), including hypotheses, milestones, and expected measurable impact\n&#8211; Experiment plans and decision memos (what was tested, results, recommendation)\n&#8211; Trained model artifacts (weights, configs, inference wrappers) with versioning and provenance\n&#8211; Benchmark suite and regression gates (scenario library, metrics, thresholds, dashboards)\n&#8211; Evaluation datasets and dataset documentation (data cards, dataset slices, known limitations)\n&#8211; Simulation scenarios and assets (procedural generation scripts, randomized environment configs)\n&#8211; Baseline implementations and reproducible training code (with unit tests where applicable)\n&#8211; Failure mode taxonomy and mitigation proposals (root causes, prioritized fixes)<\/p>\n\n\n\n<p><strong>Engineering and operational deliverables<\/strong>\n&#8211; Integration-ready modules or services (APIs, packaging, runtime dependencies)\n&#8211; Performance reports: latency\/throughput, memory footprint, robustness under distribution shift\n&#8211; MLOps-ready pipelines (training\/evaluation orchestration, experiment tracking, artifact registry)\n&#8211; Safety\/validation evidence inputs (context-specific): coverage reports, edge-case analyses\n&#8211; Documentation: model cards, integration guides, runbooks for evaluation pipelines\n&#8211; Internal training artifacts: playbooks on experiment hygiene, sim2real best practices<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">6) Goals, Objectives, and Milestones<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">30-day goals (onboarding and alignment)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Understand product goals, autonomy stack architecture, and current performance bottlenecks.<\/li>\n<li>Gain access to datasets, simulation environments, compute infrastructure, and experiment tracking.<\/li>\n<li>Reproduce one or two baseline results end-to-end (training + evaluation) to validate environment parity.<\/li>\n<li>Identify top 3\u20135 research opportunities with a clear \u201cwhy now\u201d and measurable success criteria.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">60-day goals (first contributions)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Deliver first improvement proposal with quantified impact (or a clear negative result that prevents wasted effort).<\/li>\n<li>Establish or improve a key evaluation harness (e.g., automated regression on top 50 scenarios).<\/li>\n<li>Create a failure analysis report on a high-priority issue (e.g., brittle manipulation grasping in clutter).<\/li>\n<li>Align with engineering on integration requirements: runtime constraints, API boundaries, deployment schedule.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">90-day goals (visible impact + operational maturity)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Ship a research prototype that is integration-ready (clean interfaces, reproducible training, documented evaluation).<\/li>\n<li>Demonstrate measurable performance gains on at least one primary KPI (e.g., +X% success rate on benchmark).<\/li>\n<li>Implement reproducibility improvements: standardized configs, artifact storage, seeded pipelines, tracked datasets.<\/li>\n<li>Present a 6\u201312 month research roadmap with dependencies, risks, and recommended investments.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6-month milestones<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Own a major research initiative (e.g., sim2real robustness program; foundation-model policy adaptation).<\/li>\n<li>Expand benchmark coverage and add automated regression gating used by engineering releases.<\/li>\n<li>Transfer at least one method into production or a product-grade SDK component (depending on company model).<\/li>\n<li>Demonstrate robust improvements across distributions, not just one benchmark slice (generalization proof).<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">12-month objectives<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Deliver multiple measurable autonomy capability improvements that map to product value (customer outcomes, reliability).<\/li>\n<li>Establish a sustainable research-to-product pipeline: repeatable, documented, and used by multiple teams.<\/li>\n<li>Create IP assets: patents, novel datasets, or proprietary evaluation techniques.<\/li>\n<li>Act as a recognized technical leader: mentoring, raising evaluation standards, influencing architecture decisions.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Long-term impact goals (2\u20133 years)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Enable new robotics capabilities (e.g., generalized manipulation, robust navigation in dynamic human environments).<\/li>\n<li>Reduce cost and time-to-deploy by increasing simulation validity and decreasing field trial iteration cycles.<\/li>\n<li>Build a durable competitive moat through data, evaluation, and robust autonomy algorithms.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Role success definition<\/h3>\n\n\n\n<p>Success is defined by <strong>measurable improvements in autonomy performance and robustness<\/strong>, delivered through <strong>reproducible, integration-ready research outputs<\/strong>, and adopted by engineering and product teams.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">What high performance looks like<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Consistently produces clear, decision-ready results (positive or negative) with rigorous methodology.<\/li>\n<li>Improves benchmark quality and prevents regressions through automation and gating.<\/li>\n<li>Bridges research and engineering: prototypes become maintainable modules with clear ownership boundaries.<\/li>\n<li>Anticipates future needs (compute scaling, foundation models, simulation fidelity) and prepares the org accordingly.<\/li>\n<li>Mentors others and elevates standards without becoming a bottleneck.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">7) KPIs and Productivity Metrics<\/h2>\n\n\n\n<p>The metrics below are designed to balance research discovery with product impact and operational excellence. Targets vary by company maturity, autonomy domain, and hardware constraints; example targets are illustrative.<\/p>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Metric name<\/th>\n<th>What it measures<\/th>\n<th>Why it matters<\/th>\n<th>Example target\/benchmark<\/th>\n<th>Frequency<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>Benchmark task success rate<\/td>\n<td>% of scenarios completed successfully (per task class)<\/td>\n<td>Direct measure of autonomy capability<\/td>\n<td>+5\u201315% relative improvement over baseline in 1\u20132 quarters<\/td>\n<td>Weekly\/Monthly<\/td>\n<\/tr>\n<tr>\n<td>Safety\/constraint violations rate (context-specific)<\/td>\n<td>Count\/rate of collisions, unsafe states, policy constraint breaches<\/td>\n<td>Prevents harmful behaviors; critical for deployment<\/td>\n<td>Reduce violations by 20\u201350% on targeted suite<\/td>\n<td>Weekly\/Monthly<\/td>\n<\/tr>\n<tr>\n<td>Generalization score<\/td>\n<td>Performance on held-out environments or domain-shifted scenarios<\/td>\n<td>Avoids overfitting to benchmark<\/td>\n<td>Maintain within \u22645\u201310% of in-domain performance<\/td>\n<td>Monthly<\/td>\n<\/tr>\n<tr>\n<td>Regression gate pass rate<\/td>\n<td>% of builds\/models passing evaluation thresholds<\/td>\n<td>Prevents shipping regressions<\/td>\n<td>\u226595% pass rate after stabilization<\/td>\n<td>Per release \/ Weekly<\/td>\n<\/tr>\n<tr>\n<td>Experiment cycle time<\/td>\n<td>Time from hypothesis \u2192 result \u2192 decision<\/td>\n<td>Measures iteration velocity<\/td>\n<td>Median cycle \u22647\u201314 days for standard experiments<\/td>\n<td>Weekly<\/td>\n<\/tr>\n<tr>\n<td>Reproducibility rate<\/td>\n<td>% of key results reproducible by others or reruns<\/td>\n<td>Reduces \u201cfragile wins\u201d<\/td>\n<td>\u226590% reproducibility for reported improvements<\/td>\n<td>Monthly<\/td>\n<\/tr>\n<tr>\n<td>Compute efficiency<\/td>\n<td>Performance gain per GPU-hour or per training run<\/td>\n<td>Controls cost, improves sustainability<\/td>\n<td>10\u201330% cost reduction for equivalent performance<\/td>\n<td>Monthly\/Quarterly<\/td>\n<\/tr>\n<tr>\n<td>Sim2real transfer delta<\/td>\n<td>Difference between simulation performance and real-world\/target-domain performance<\/td>\n<td>Indicates deployment readiness<\/td>\n<td>Reduce sim2real gap by 20\u201340% on key tasks<\/td>\n<td>Quarterly<\/td>\n<\/tr>\n<tr>\n<td>Failure mode recurrence<\/td>\n<td>Rate of repeated known failures after mitigation<\/td>\n<td>Measures effectiveness of fixes<\/td>\n<td>\u226550% reduction in top 3 failure modes<\/td>\n<td>Monthly<\/td>\n<\/tr>\n<tr>\n<td>Model runtime latency<\/td>\n<td>Inference time under target hardware constraints<\/td>\n<td>Ensures deployability<\/td>\n<td>Meet SLA (e.g., &lt;30ms per frame)<\/td>\n<td>Per build<\/td>\n<\/tr>\n<tr>\n<td>Memory\/footprint<\/td>\n<td>RAM\/VRAM and storage requirements<\/td>\n<td>Impacts embedded deployment viability<\/td>\n<td>Within platform budget (e.g., &lt;2GB VRAM)<\/td>\n<td>Per build<\/td>\n<\/tr>\n<tr>\n<td>Data coverage score<\/td>\n<td>Coverage of scenario factors (lighting, clutter, dynamics) in training\/eval<\/td>\n<td>Improves robustness<\/td>\n<td>Coverage increase of key factors quarter-over-quarter<\/td>\n<td>Quarterly<\/td>\n<\/tr>\n<tr>\n<td>Adoption rate of research outputs<\/td>\n<td># of research deliverables used in product or by other teams<\/td>\n<td>Ensures transfer and impact<\/td>\n<td>1\u20133 adopted components per year (varies by org)<\/td>\n<td>Quarterly<\/td>\n<\/tr>\n<tr>\n<td>Stakeholder satisfaction<\/td>\n<td>Partner feedback (engineering\/product) on clarity and usefulness<\/td>\n<td>Ensures collaboration effectiveness<\/td>\n<td>\u22654.2\/5 average internal survey<\/td>\n<td>Quarterly<\/td>\n<\/tr>\n<tr>\n<td>Mentorship impact (Senior IC)<\/td>\n<td>Growth of junior staff, quality of reviews, knowledge sharing<\/td>\n<td>Scales team capability<\/td>\n<td>Regular mentoring + documented playbooks<\/td>\n<td>Quarterly<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">8) Technical Skills Required<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Must-have technical skills<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Robotics ML fundamentals (Critical)<\/strong> <\/li>\n<li>Description: Core ML methods used in robotics: supervised learning for perception, sequential decision-making, policy learning, representation learning.  <\/li>\n<li>Use: Designing and training models for perception\/policy; selecting appropriate objectives and evaluation.  <\/li>\n<li><strong>Experimentation rigor and statistics (Critical)<\/strong> <\/li>\n<li>Description: Ablations, baselines, significance thinking, dataset splits, leakage prevention.  <\/li>\n<li>Use: Producing trustworthy conclusions and decision memos.  <\/li>\n<li><strong>Deep learning frameworks: PyTorch (Critical) and\/or JAX (Important)<\/strong> <\/li>\n<li>Description: Training pipelines, custom losses, distributed training basics.  <\/li>\n<li>Use: Implementing and iterating on models and policies.  <\/li>\n<li><strong>Robotics concepts: kinematics, dynamics, control, estimation (Critical)<\/strong> <\/li>\n<li>Description: Understanding robot motion, sensors, feedback control, state estimation.  <\/li>\n<li>Use: Designing learning systems that respect physical constraints and integrate with classical stacks.  <\/li>\n<li><strong>3D perception and sensor processing (Important to Critical depending on focus)<\/strong> <\/li>\n<li>Description: Working with RGB-D, LiDAR, IMU; point clouds; tracking; coordinate frames.  <\/li>\n<li>Use: Perception modules and state representations for downstream planning\/control.  <\/li>\n<li><strong>Software engineering in Python (Critical) + C++ familiarity (Important)<\/strong> <\/li>\n<li>Description: Clean code, modular design, profiling, debugging; ability to interface with C++ stacks.  <\/li>\n<li>Use: Building research code that can be handed off to engineering and integrated.  <\/li>\n<li><strong>Simulation-driven development (Important)<\/strong> <\/li>\n<li>Description: Using simulators, scenario generation, sensor models, domain randomization.  <\/li>\n<li>Use: Scaling experiments and evaluating long-tail scenarios.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Good-to-have technical skills<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>ROS2 ecosystem (Important, Common in robotics)<\/strong> <\/li>\n<li>Use: Integration with robotics middleware, message definitions, nodes, and runtime tools.  <\/li>\n<li><strong>Reinforcement learning and imitation learning (Important)<\/strong> <\/li>\n<li>Use: Learning control policies, manipulation strategies, navigation behaviors.  <\/li>\n<li><strong>Motion planning familiarity (Optional to Important)<\/strong> <\/li>\n<li>Use: Hybrid systems combining learning with sampling-based or optimization-based planners.  <\/li>\n<li><strong>Multi-modal learning (Important)<\/strong> <\/li>\n<li>Use: Vision + language + proprioception fusion for instruction-following or semantic generalization.  <\/li>\n<li><strong>Distributed training and GPU scaling (Important)<\/strong> <\/li>\n<li>Use: Efficient use of multi-GPU, cluster scheduling, experiment parallelism.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Advanced or expert-level technical skills<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Sim2real transfer expertise (Critical for many robotics products)<\/strong> <\/li>\n<li>Use: Domain randomization design, residual learning, system identification, adaptation pipelines.  <\/li>\n<li><strong>Model-based RL \/ world models (Optional to Important, Emerging)<\/strong> <\/li>\n<li>Use: Sample-efficient learning, planning in latent space, long-horizon control.  <\/li>\n<li><strong>Uncertainty estimation and robustness methods (Important)<\/strong> <\/li>\n<li>Use: OOD detection, calibration, risk-aware decision making, safe fallback behaviors.  <\/li>\n<li><strong>Optimization for deployment (Important)<\/strong> <\/li>\n<li>Use: Quantization, pruning, distillation, TensorRT\/ONNX optimization; latency profiling.  <\/li>\n<li><strong>Advanced evaluation design (Critical for Senior level)<\/strong> <\/li>\n<li>Use: Scenario factorization, stress testing, coverage metrics, regression gate design.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Emerging future skills for this role (next 2\u20135 years)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Vision-Language-Action (VLA) and foundation model adaptation (Important, Emerging)<\/strong> <\/li>\n<li>Use: Fine-tuning or adapting large models to robotics tasks with small data, constraints, and safety layers.  <\/li>\n<li><strong>Synthetic data at scale + procedural world generation (Important, Emerging)<\/strong> <\/li>\n<li>Use: Improving generalization through diverse training distributions and automated labeling.  <\/li>\n<li><strong>Automated evaluation and verification tooling (Important, Emerging)<\/strong> <\/li>\n<li>Use: Continuous autonomy evaluation (CAE), property testing for policies, formal-ish checks in constrained domains.  <\/li>\n<li><strong>On-robot continual learning governance (Optional, Context-specific)<\/strong> <\/li>\n<li>Use: Safe update mechanisms, drift detection, and controlled adaptation in deployed fleets.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">9) Soft Skills and Behavioral Capabilities<\/h2>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Scientific judgment and intellectual honesty<\/strong> <\/li>\n<li>Why it matters: Robotics research is prone to overfitting and fragile wins; integrity protects the business from false confidence.  <\/li>\n<li>Shows up as: Clear baselines, transparent limitations, negative results shared early.  <\/li>\n<li>\n<p>Strong performance: Makes decisions based on evidence; prevents wasted quarters on unscalable approaches.<\/p>\n<\/li>\n<li>\n<p><strong>Systems thinking (research-to-product mindset)<\/strong> <\/p>\n<\/li>\n<li>Why it matters: Robotics performance depends on interactions between perception, planning, control, runtime, and data.  <\/li>\n<li>Shows up as: Proposes solutions that consider integration constraints, compute budgets, and observability.  <\/li>\n<li>\n<p>Strong performance: Delivers improvements that survive integration and real-world variability.<\/p>\n<\/li>\n<li>\n<p><strong>Technical communication (multi-audience)<\/strong> <\/p>\n<\/li>\n<li>Why it matters: Stakeholders include product, engineering, leadership; clarity accelerates decisions.  <\/li>\n<li>Shows up as: Decision memos, crisp tradeoff presentations, well-structured docs.  <\/li>\n<li>\n<p>Strong performance: Stakeholders can act immediately based on outputs.<\/p>\n<\/li>\n<li>\n<p><strong>Collaboration and influence without authority<\/strong> <\/p>\n<\/li>\n<li>Why it matters: Senior ICs must align cross-functional teams on evaluation standards and integration choices.  <\/li>\n<li>Shows up as: Negotiates scope, sets shared metrics, resolves conflicts constructively.  <\/li>\n<li>\n<p>Strong performance: Teams adopt benchmarks and methods voluntarily because they see value.<\/p>\n<\/li>\n<li>\n<p><strong>Prioritization under uncertainty<\/strong> <\/p>\n<\/li>\n<li>Why it matters: Many research directions are plausible; compute\/time is limited.  <\/li>\n<li>Shows up as: Focuses on highest expected value experiments; avoids open-ended exploration without gates.  <\/li>\n<li>\n<p>Strong performance: Produces a steady stream of decision points and de-risks programs early.<\/p>\n<\/li>\n<li>\n<p><strong>Mentorship and talent multiplier behavior (Senior expectation)<\/strong> <\/p>\n<\/li>\n<li>Why it matters: Research quality improves when best practices are shared; prevents single points of failure.  <\/li>\n<li>Shows up as: Coaching on experiment hygiene, review habits, reproducibility frameworks.  <\/li>\n<li>\n<p>Strong performance: Junior staff ship stronger work faster; fewer repeated mistakes.<\/p>\n<\/li>\n<li>\n<p><strong>Resilience and iterative problem-solving<\/strong> <\/p>\n<\/li>\n<li>Why it matters: Robotics failure rates can be high; progress is nonlinear.  <\/li>\n<li>Shows up as: Calm debugging, disciplined iteration, avoids blame.  <\/li>\n<li>Strong performance: Converts setbacks into learning and improved evaluation coverage.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">10) Tools, Platforms, and Software<\/h2>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Category<\/th>\n<th>Tool \/ platform \/ software<\/th>\n<th>Primary use<\/th>\n<th>Common \/ Optional \/ Context-specific<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>Cloud platforms<\/td>\n<td>AWS, GCP, Azure<\/td>\n<td>GPU compute, storage, managed services<\/td>\n<td>Context-specific (varies by company)<\/td>\n<\/tr>\n<tr>\n<td>Compute orchestration<\/td>\n<td>Kubernetes<\/td>\n<td>Scheduling training\/eval workloads<\/td>\n<td>Common (in larger orgs)<\/td>\n<\/tr>\n<tr>\n<td>ML experiment tracking<\/td>\n<td>MLflow, Weights &amp; Biases<\/td>\n<td>Track runs, metrics, artifacts<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Data\/versioning<\/td>\n<td>DVC, LakeFS, Git LFS<\/td>\n<td>Dataset versioning and provenance<\/td>\n<td>Optional (org-dependent)<\/td>\n<\/tr>\n<tr>\n<td>Data processing<\/td>\n<td>Spark, Ray<\/td>\n<td>Large-scale preprocessing, distributed workloads<\/td>\n<td>Optional<\/td>\n<\/tr>\n<tr>\n<td>AI\/ML frameworks<\/td>\n<td>PyTorch; JAX<\/td>\n<td>Training models\/policies<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Model optimization<\/td>\n<td>ONNX, TensorRT<\/td>\n<td>Inference optimization<\/td>\n<td>Context-specific (deployment dependent)<\/td>\n<\/tr>\n<tr>\n<td>Robotics middleware<\/td>\n<td>ROS2<\/td>\n<td>Messaging, nodes, runtime tools<\/td>\n<td>Common (robotics orgs)<\/td>\n<\/tr>\n<tr>\n<td>Simulation platforms<\/td>\n<td>NVIDIA Isaac Sim; Gazebo\/Ignition; MuJoCo<\/td>\n<td>Simulation, synthetic data, policy evaluation<\/td>\n<td>Common (one or more)<\/td>\n<\/tr>\n<tr>\n<td>Physics \/ dynamics<\/td>\n<td>Pinocchio; Drake<\/td>\n<td>Dynamics, kinematics, planning utilities<\/td>\n<td>Optional<\/td>\n<\/tr>\n<tr>\n<td>3D processing<\/td>\n<td>Open3D; PCL<\/td>\n<td>Point cloud processing and visualization<\/td>\n<td>Optional<\/td>\n<\/tr>\n<tr>\n<td>Computer vision<\/td>\n<td>OpenCV<\/td>\n<td>Image processing, calibration utilities<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>DevOps \/ CI-CD<\/td>\n<td>GitHub Actions; GitLab CI; Jenkins<\/td>\n<td>CI pipelines for tests\/evals<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Containers<\/td>\n<td>Docker<\/td>\n<td>Reproducible environments<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Observability<\/td>\n<td>Prometheus; Grafana<\/td>\n<td>Metrics dashboards for pipelines\/services<\/td>\n<td>Optional<\/td>\n<\/tr>\n<tr>\n<td>Logging<\/td>\n<td>ELK\/Elastic; OpenTelemetry<\/td>\n<td>Log aggregation, tracing<\/td>\n<td>Optional<\/td>\n<\/tr>\n<tr>\n<td>Source control<\/td>\n<td>Git (GitHub\/GitLab\/Bitbucket)<\/td>\n<td>Code management<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>IDE \/ engineering<\/td>\n<td>VS Code; JetBrains<\/td>\n<td>Development environment<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Collaboration<\/td>\n<td>Slack\/MS Teams; Confluence\/Notion<\/td>\n<td>Communication and documentation<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Project tracking<\/td>\n<td>Jira; Linear<\/td>\n<td>Work planning and tracking<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Security (data)<\/td>\n<td>Vault; KMS<\/td>\n<td>Secrets management, encryption<\/td>\n<td>Context-specific<\/td>\n<\/tr>\n<tr>\n<td>Artifact registry<\/td>\n<td>Artifactory; S3\/GCS registries<\/td>\n<td>Store models\/build artifacts<\/td>\n<td>Common<\/td>\n<\/tr>\n<tr>\n<td>Testing \/ QA<\/td>\n<td>pytest; unit\/integration test frameworks<\/td>\n<td>Test evaluation harness and utilities<\/td>\n<td>Common<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">11) Typical Tech Stack \/ Environment<\/h2>\n\n\n\n<p><strong>Infrastructure environment<\/strong>\n&#8211; GPU-enabled compute: on-prem cluster, cloud GPUs, or hybrid\n&#8211; Containerized workloads with Docker; often scheduled via Kubernetes or batch schedulers\n&#8211; Artifact storage for models and datasets (object storage) with access controls\n&#8211; High-throughput networking and storage for large robotics datasets (images, point clouds, logs)<\/p>\n\n\n\n<p><strong>Application environment<\/strong>\n&#8211; Research codebases in Python with performance-critical components in C++ (and occasional CUDA)\n&#8211; Robotics runtime stack often based on ROS2, plus proprietary components for deployment\n&#8211; Inference services or embedded runtime (depending on robot platform): may require strict latency constraints<\/p>\n\n\n\n<p><strong>Data environment<\/strong>\n&#8211; Multi-modal datasets: video, depth, LiDAR, IMU, joint states, actions, maps\n&#8211; Labeling pipelines (internal tools or vendors) and synthetic data generation\n&#8211; Dataset governance: splits, lineage, data cards, and privacy controls (where applicable)<\/p>\n\n\n\n<p><strong>Security environment<\/strong>\n&#8211; Role-based access to datasets and model artifacts\n&#8211; Encryption at rest and in transit for sensitive logs (context-specific)\n&#8211; Compliance with internal model governance policies (e.g., model cards, risk reviews)<\/p>\n\n\n\n<p><strong>Delivery model<\/strong>\n&#8211; Research-to-product transfer model: prototypes promoted to \u201ccandidate components\u201d with defined APIs and owners\n&#8211; CI-based evaluation: automated benchmarks as gating for merges\/releases (maturity varies)<\/p>\n\n\n\n<p><strong>Agile\/SDLC context<\/strong>\n&#8211; Mix of Agile engineering rhythms (sprints, Jira) and research rhythms (milestones, quarterly roadmaps)\n&#8211; Peer review practices for research code and experiment design\n&#8211; Release coordination with product teams when research outputs land in customer-facing features<\/p>\n\n\n\n<p><strong>Scale\/complexity context<\/strong>\n&#8211; High variability environments and long-tail failures; evaluation at scale is a core differentiator\n&#8211; Compute cost management is non-trivial; experiment prioritization and efficiency matter\n&#8211; Multiple robot configurations\/sensors may exist, increasing combinatorial complexity<\/p>\n\n\n\n<p><strong>Team topology<\/strong>\n&#8211; Senior Robotics Research Scientist typically sits in an AI &amp; ML org with dotted-line collaboration to Robotics Engineering\n&#8211; Common structure: small research pod (2\u20136) + applied ML engineers + platform\/simulation team + product owner<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">12) Stakeholders and Collaboration Map<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Internal stakeholders<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Head\/Director of Robotics Research \/ Director of AI Research (manager)<\/strong> <\/li>\n<li>Collaboration: roadmap alignment, priorities, compute budget, staffing needs  <\/li>\n<li>Escalation: strategic tradeoffs, timeline conflicts, cross-org dependencies<\/li>\n<li><strong>Robotics Software Engineering Lead<\/strong> <\/li>\n<li>Collaboration: integration, runtime constraints, API contracts, release readiness  <\/li>\n<li>Escalation: production feasibility, performance regressions, reliability concerns<\/li>\n<li><strong>Applied ML \/ MLOps team<\/strong> <\/li>\n<li>Collaboration: training pipelines, deployment patterns, model registry, monitoring  <\/li>\n<li>Escalation: pipeline instability, reproducibility gaps, platform limitations<\/li>\n<li><strong>Simulation\/Infrastructure team<\/strong> <\/li>\n<li>Collaboration: scenario generation, simulator fidelity, synthetic data pipelines  <\/li>\n<li>Escalation: simulator bugs affecting benchmark validity, compute bottlenecks<\/li>\n<li><strong>Product Management (robotics\/autonomy)<\/strong> <\/li>\n<li>Collaboration: success metrics tied to customer value, feature prioritization  <\/li>\n<li>Escalation: mismatch between research goals and product timeline\/market needs<\/li>\n<li><strong>QA\/Validation\/Safety (context-specific)<\/strong> <\/li>\n<li>Collaboration: test coverage, acceptance criteria, evidence collection  <\/li>\n<li>Escalation: safety-critical regressions, insufficient validation rigor<\/li>\n<li><strong>Data Engineering \/ Data Ops<\/strong> <\/li>\n<li>Collaboration: ingestion pipelines, schema, storage, labeling, governance  <\/li>\n<li>Escalation: data quality gaps, label inconsistency, missing provenance<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">External stakeholders (as applicable)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Academic\/industry partners<\/strong>: joint research, sponsored projects, internships (policy-dependent)<\/li>\n<li><strong>Vendors<\/strong>: labeling vendors, simulation tool vendors, hardware platform vendors<\/li>\n<li><strong>Customers \/ field partners<\/strong>: feedback loops, telemetry review, domain constraints<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Peer roles<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Senior\/Staff Applied Scientist (ML), Robotics Engineer, Simulation Engineer, MLOps Engineer, Product Manager<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Upstream dependencies<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Sensor\/log data availability, data labeling throughput, simulator updates, compute capacity, platform stability<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Downstream consumers<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Robotics autonomy stack, SDKs, product feature teams, evaluation\/QA gating systems, customer deployment teams<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Nature of collaboration<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Highly iterative, evidence-driven, and dependency-heavy; success requires alignment on metrics and acceptance thresholds.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Typical decision-making authority<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Owns technical research decisions within scope (methods, experiments, benchmarks)  <\/li>\n<li>Shares decisions on integration design with engineering  <\/li>\n<li>Escalates product tradeoffs and major architecture changes to leadership<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Escalation points<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Safety or deployment risks<\/li>\n<li>Major regressions or benchmark validity concerns<\/li>\n<li>Significant compute budget increases<\/li>\n<li>Cross-team priority conflicts<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">13) Decision Rights and Scope of Authority<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Can decide independently<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Experiment design, baselines, ablation plans, and evaluation methodology within agreed standards<\/li>\n<li>Selection of modeling approaches for prototypes (within platform constraints)<\/li>\n<li>Dataset slicing strategies for analysis (using approved data sources)<\/li>\n<li>Research code structure and internal tooling improvements<\/li>\n<li>Publication proposals (subject to company review process)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Requires team approval (research\/engineering consensus)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Changes to benchmark definitions and regression thresholds that gate releases<\/li>\n<li>Adoption of new simulation frameworks or major scenario library changes<\/li>\n<li>Integration API designs that affect multiple components<\/li>\n<li>Modifications to shared datasets, labeling guidelines, or data schemas<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Requires manager\/director approval<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Compute budget increases beyond agreed allocation<\/li>\n<li>Shifts in research roadmap priorities with material product impact<\/li>\n<li>External collaborations, open-sourcing, or conference submissions (per policy)<\/li>\n<li>Hiring requests for interns, contractors, or additional headcount<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Requires executive approval (VP\/CTO-level, typical)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Major platform bets (new autonomy stack direction, large-scale foundation model program)<\/li>\n<li>Vendor contracts with material spend<\/li>\n<li>Strategic IP actions (patent portfolio direction, acquisitions\/partnerships)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Budget\/architecture\/vendor authority<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Budget: typically influences compute spend and vendor recommendations; final approval rests with leadership<\/li>\n<li>Architecture: can propose and drive adoption of algorithms and evaluation architecture; production architecture final sign-off often shared with Engineering leadership<\/li>\n<li>Vendors: can evaluate tools; procurement decisions follow standard enterprise processes<\/li>\n<li>Hiring: participates in interviews and hiring decisions; final approval by manager and HR process<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">14) Required Experience and Qualifications<\/h2>\n\n\n\n<p><strong>Typical years of experience<\/strong>\n&#8211; Usually <strong>6\u201310+ years<\/strong> in robotics research, applied ML, or autonomy-related development (industry and\/or PhD research), with evidence of end-to-end ownership.<\/p>\n\n\n\n<p><strong>Education expectations<\/strong>\n&#8211; <strong>PhD<\/strong> in Robotics, Computer Science, Electrical Engineering, Mechanical Engineering, or related field is common and often preferred for \u201cSenior Research Scientist\u201d tracks.<br\/>\n&#8211; Strong candidates may have an MS + significant industry research impact and publication\/patent record.<\/p>\n\n\n\n<p><strong>Certifications (generally not primary for this role)<\/strong>\n&#8211; Not typically required.<br\/>\n&#8211; Optional\/context-specific: cloud certifications (AWS\/GCP) if heavily infrastructure-oriented; safety certifications if in regulated robotics contexts.<\/p>\n\n\n\n<p><strong>Prior role backgrounds commonly seen<\/strong>\n&#8211; Robotics Research Scientist, Applied Scientist (Robotics\/Autonomy), ML Researcher (embodied AI), Robotics Software Engineer with strong ML focus\n&#8211; Academic postdoc or research engineer in robotics labs transitioning to industry<\/p>\n\n\n\n<p><strong>Domain knowledge expectations<\/strong>\n&#8211; Robotics autonomy domain knowledge: perception + decision-making + control interplay\n&#8211; Practical experience with simulation and real-world data issues (calibration, noise, drift)\n&#8211; Comfort with constraints and tradeoffs: latency, compute, reliability, maintainability<\/p>\n\n\n\n<p><strong>Leadership experience expectations (Senior IC)<\/strong>\n&#8211; Demonstrated leadership through technical direction, mentorship, and cross-functional delivery\n&#8211; Not required to be a people manager, but should show ability to lead initiatives and influence standards<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">15) Career Path and Progression<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Common feeder roles into this role<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Robotics Research Scientist (mid-level)<\/li>\n<li>Applied Scientist \/ ML Scientist focused on autonomy or embodied AI<\/li>\n<li>Senior Robotics Engineer with research-grade experimentation experience<\/li>\n<li>Research Engineer in simulation\/synthetic data with ML depth<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Next likely roles after this role<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Staff Robotics Research Scientist<\/strong> (broader scope, cross-domain ownership, benchmark governance)<\/li>\n<li><strong>Principal Robotics Research Scientist<\/strong> (company-wide autonomy strategy, IP leadership)<\/li>\n<li><strong>Tech Lead \/ Staff Applied Scientist (Robotics)<\/strong> bridging research and production across multiple teams<\/li>\n<li><strong>Robotics Research Manager<\/strong> (if moving into people leadership)<\/li>\n<li><strong>Autonomy Architect \/ Robotics AI Architect<\/strong> (systems-level design authority)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Adjacent career paths<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Simulation &amp; Synthetic Data Lead<\/strong> (focus on scalable scenario generation and validation)<\/li>\n<li><strong>MLOps for Robotics<\/strong> (production training\/eval pipelines and monitoring)<\/li>\n<li><strong>Robotics Product\/Strategy<\/strong> (for those strong in market translation, less hands-on research)<\/li>\n<li><strong>Safety\/Validation specialist (context-specific)<\/strong> in regulated robotics applications<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Skills needed for promotion (Senior \u2192 Staff)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Ownership of multi-quarter programs with clear outcomes and adoption<\/li>\n<li>Standard-setting: benchmarks, evaluation governance, best practices<\/li>\n<li>Ability to unify research and engineering execution across teams<\/li>\n<li>Track record of shipping research outcomes into product-grade systems<\/li>\n<li>Strong mentorship and multiplication of team capability<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">How this role evolves over time<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>From \u201cdelivering improvements to a component\u201d \u2192 \u201cowning a capability area end-to-end\u201d  <\/li>\n<li>From \u201crunning experiments\u201d \u2192 \u201cbuilding the evaluation and data engines that scale experiments org-wide\u201d  <\/li>\n<li>From \u201cmodel performance\u201d \u2192 \u201crobustness, safety, generalization, and cost-to-deploy as first-class outcomes\u201d<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">16) Risks, Challenges, and Failure Modes<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Common role challenges<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Sim-to-real gap<\/strong>: strong simulation results that fail in the physical world due to modeling inaccuracies or distribution shift.<\/li>\n<li><strong>Benchmark overfitting<\/strong>: improvements that do not generalize beyond curated scenarios.<\/li>\n<li><strong>Data constraints<\/strong>: insufficient diversity, label noise, missing edge cases, or privacy restrictions.<\/li>\n<li><strong>Integration friction<\/strong>: research prototypes that are hard to productionize due to dependencies, unclear APIs, or performance constraints.<\/li>\n<li><strong>Compute limitations<\/strong>: inability to run adequate ablations or scale training; poor prioritization wastes budget.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Bottlenecks<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Slow labeling throughput or ambiguous labeling specs<\/li>\n<li>Simulator instability or limited fidelity<\/li>\n<li>Fragmented telemetry and insufficient observability in deployed systems<\/li>\n<li>Lack of standardized evaluation; too many bespoke scripts<\/li>\n<li>Cross-team misalignment on success metrics and acceptance thresholds<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Anti-patterns<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>\u201cOne-run wins\u201d presented as breakthroughs without reproducibility<\/li>\n<li>Changing multiple variables at once (no ablations), making progress non-attributable<\/li>\n<li>Neglecting baselines and simple approaches; jumping to complex models prematurely<\/li>\n<li>Producing research that cannot be integrated (no documentation, no tests, unclear dependencies)<\/li>\n<li>Treating evaluation as an afterthought rather than a product<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Common reasons for underperformance<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Weak experimental rigor or inability to interpret results<\/li>\n<li>Overemphasis on novelty over measurable product impact<\/li>\n<li>Poor cross-functional collaboration; inability to translate constraints<\/li>\n<li>Lack of ownership: many ideas, few completed deliverables adopted by others<\/li>\n<li>Inadequate coding practices leading to unmaintainable research artifacts<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Business risks if this role is ineffective<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Autonomy regressions or unreliable robotics performance leading to customer churn<\/li>\n<li>Increased deployment costs due to excessive field testing and manual tuning<\/li>\n<li>Missed market windows due to slow iteration and unclear technical direction<\/li>\n<li>Wasted compute and engineering time on unproductive research paths<\/li>\n<li>Reduced trust in AI outputs and internal evaluation processes<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">17) Role Variants<\/h2>\n\n\n\n<p>This role is consistent in its core (robotics AI research + evaluation + transfer), but scope and emphasis vary.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">By company size<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Startup (early-stage robotics software)<\/strong> <\/li>\n<li>Broader scope: research + engineering + some productization; faster prototyping; less platform maturity.  <\/li>\n<li>Higher tolerance for tech debt initially, but senior scientist must still enforce minimal reproducibility.<\/li>\n<li><strong>Mid-size scale-up<\/strong> <\/li>\n<li>Balanced: dedicated simulation\/data teams exist; clearer roadmaps; stronger handoff expectations.  <\/li>\n<li>More formal benchmarks and CI gating begin to emerge.<\/li>\n<li><strong>Enterprise<\/strong> <\/li>\n<li>Strong governance: model risk reviews, security, procurement, QA.  <\/li>\n<li>Longer integration cycles; high emphasis on documentation, reliability, and cross-team alignment.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">By industry<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>General software\/IT robotics platform provider<\/strong> (broad applicability)  <\/li>\n<li>Focus on SDKs, simulation tools, autonomy modules as reusable software products.<\/li>\n<li><strong>Warehouse\/industrial automation (context-specific)<\/strong> <\/li>\n<li>Strong emphasis on safety, uptime, and operational metrics; constrained environments but high reliability needs.<\/li>\n<li><strong>Consumer robotics (context-specific)<\/strong> <\/li>\n<li>Emphasis on cost\/compute constraints, user experience, privacy, and robustness in messy environments.<\/li>\n<li><strong>Healthcare or regulated domains (context-specific)<\/strong> <\/li>\n<li>Strong validation evidence requirements, traceability, and risk management.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">By geography<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Core skills are global; differences appear in:<\/li>\n<li>Data handling and privacy requirements<\/li>\n<li>Export controls or compute access constraints (context-specific)<\/li>\n<li>Talent market norms (PhD prevalence, publication expectations)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Product-led vs service-led company<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Product-led<\/strong> <\/li>\n<li>Clear product milestones, roadmap alignment, and reusable components; stronger pressure for integration-ready outputs.<\/li>\n<li><strong>Service-led \/ solutions<\/strong> <\/li>\n<li>More customization per customer environment; evaluation must handle varied deployments; greater on-site\/field feedback loops.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Startup vs enterprise operating model<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Startup<\/strong>: rapid iteration; the Senior Scientist may directly own deployment experiments.  <\/li>\n<li><strong>Enterprise<\/strong>: stricter release gating, specialized teams; the Senior Scientist influences through standards and cross-functional programs.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Regulated vs non-regulated environment<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Regulated\/safety-critical<\/strong> <\/li>\n<li>Higher emphasis on traceability, test coverage, and documented evidence.  <\/li>\n<li>Additional collaboration with safety\/quality and compliance functions.<\/li>\n<li><strong>Non-regulated<\/strong> <\/li>\n<li>Faster experimentation; still requires internal governance to prevent reliability failures.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">18) AI \/ Automation Impact on the Role<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Tasks that can be automated (now and increasing)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Literature triage and summarization<\/strong>: automated tools can summarize papers, compare methods, and generate reading lists (human validation required).<\/li>\n<li><strong>Experiment boilerplate<\/strong>: template generation for training scripts, configs, and evaluation harness scaffolding.<\/li>\n<li><strong>Hyperparameter sweeps and search<\/strong>: automated tuning, scheduling, and early stopping.<\/li>\n<li><strong>Dataset quality checks<\/strong>: automated detection of label anomalies, duplicates, corrupted files, distribution drift.<\/li>\n<li><strong>Regression detection<\/strong>: automated alerts when benchmark KPIs drop below thresholds.<\/li>\n<li><strong>Synthetic data generation<\/strong>: procedural scene generation and auto-labeling pipelines in simulation.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Tasks that remain human-critical<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem selection and research judgment<\/strong>: deciding what matters for product impact and feasibility under constraints.<\/li>\n<li><strong>Causal reasoning and failure interpretation<\/strong>: understanding why policies fail in certain conditions and what mitigations are safe.<\/li>\n<li><strong>Evaluation design<\/strong>: selecting scenario factors, preventing leakage, and ensuring metrics reflect real-world value.<\/li>\n<li><strong>Cross-functional influence<\/strong>: aligning product\/engineering\/safety on tradeoffs and adoption.<\/li>\n<li><strong>Ethical\/safety governance (context-specific)<\/strong>: deciding appropriate constraints, update policies, and risk controls.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">How AI changes the role over the next 2\u20135 years<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Shift from model-building to system-building<\/strong>: more emphasis on orchestrating foundation models, adapters, retrieval, and constraint layers rather than training from scratch.<\/li>\n<li><strong>Continuous autonomy evaluation becomes standard<\/strong>: always-on benchmark pipelines integrated into CI\/CD, with richer scenario coverage metrics.<\/li>\n<li><strong>Synthetic data becomes a primary lever<\/strong>: large-scale procedural generation and sim instrumentation will drive robustness gains.<\/li>\n<li><strong>New skill expectations<\/strong>: foundation model adaptation, prompt\/programmatic interfaces for embodied models, automated test generation for robotics behaviors.<\/li>\n<li><strong>Higher expectations for governance<\/strong>: reproducibility, lineage, and safety constraints become mandatory as autonomy systems scale.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">New expectations caused by AI, automation, or platform shifts<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Deliver \u201cevaluation as code\u201d with strong automation and traceability.<\/li>\n<li>Demonstrate robustness and calibration, not just average-case performance.<\/li>\n<li>Operate with cost discipline: optimize for GPU spend and iteration velocity.<\/li>\n<li>Maintain higher security standards around data and model artifacts (especially with multi-modal telemetry).<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">19) Hiring Evaluation Criteria<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">What to assess in interviews<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Robotics + ML depth<\/strong>: Can the candidate reason across perception\/policy\/control and understand physical constraints?<\/li>\n<li><strong>Experimentation rigor<\/strong>: Do they design clean ablations, avoid leakage, and communicate uncertainty?<\/li>\n<li><strong>Systems and integration mindset<\/strong>: Can they translate research into components that engineering can ship?<\/li>\n<li><strong>Simulation and evaluation expertise<\/strong>: Can they create scalable, meaningful benchmarks?<\/li>\n<li><strong>Pragmatism under constraints<\/strong>: Can they choose methods that fit latency\/compute\/data realities?<\/li>\n<li><strong>Collaboration and leadership as Senior IC<\/strong>: Can they mentor, influence, and lead initiatives without formal authority?<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">Practical exercises or case studies (recommended)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Case study: autonomy failure analysis<\/strong> <\/li>\n<li>Provide logs\/metrics and a description of failure (e.g., navigation fails in reflective floors; manipulation slips in clutter).  <\/li>\n<li>Ask for root-cause hypotheses, proposed experiments, and evaluation changes.<\/li>\n<li><strong>Design exercise: benchmark and gating plan<\/strong> <\/li>\n<li>Ask candidate to propose metrics, scenario factors, and regression gating for a new capability.<\/li>\n<li><strong>Research transfer exercise<\/strong> <\/li>\n<li>Present a research prototype concept and ask how they would productionize it: APIs, tests, monitoring, model registry, rollout strategy.<\/li>\n<li><strong>Technical deep dive<\/strong> <\/li>\n<li>Candidate presents 1\u20132 projects with details: data, models, evaluation, failures, and what they\u2019d do differently.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Strong candidate signals<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Demonstrated end-to-end ownership: from hypothesis to measurable improvement to integration\/adoption.<\/li>\n<li>Clear thinking about generalization, distribution shift, and sim2real transfer.<\/li>\n<li>Evidence of reproducibility discipline (tracked experiments, versioned datasets, stable baselines).<\/li>\n<li>Ability to articulate tradeoffs and decision criteria (not just \u201cstate-of-the-art\u201d claims).<\/li>\n<li>Mentorship examples and standard-setting behaviors (benchmarking, code quality, review culture).<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Weak candidate signals<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Focus on novelty without metrics tied to real outcomes.<\/li>\n<li>Vague evaluation; no baselines; no ablations.<\/li>\n<li>Inability to discuss failures or negative results.<\/li>\n<li>\u201cIt worked on my machine\u201d mentality; weak engineering hygiene.<\/li>\n<li>Overreliance on black-box methods without interpretability or safety considerations.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Red flags<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Claims large gains without being able to explain evaluation design or data splits.<\/li>\n<li>Dismisses integration constraints (latency, memory, runtime determinism).<\/li>\n<li>Poor collaboration posture (\u201cengineering should just integrate it\u201d).<\/li>\n<li>Does not consider safety\/robustness in robotics contexts.<\/li>\n<li>Repeatedly conflates benchmark performance with deployment readiness.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Scorecard dimensions (interview rubric)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Robotics fundamentals (kinematics\/dynamics\/control)<\/li>\n<li>ML depth (representation learning, sequential decision-making, multi-modal)<\/li>\n<li>Experiment design and statistical thinking<\/li>\n<li>Simulation, evaluation, and benchmarking rigor<\/li>\n<li>Software engineering quality (readability, testing mindset, performance awareness)<\/li>\n<li>Research-to-product transfer mindset<\/li>\n<li>Communication clarity (written + verbal)<\/li>\n<li>Collaboration and leadership (mentorship, influence)<\/li>\n<li>Domain adaptability and learning agility<\/li>\n<li>Integrity and scientific judgment<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">20) Final Role Scorecard Summary<\/h2>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Category<\/th>\n<th>Summary<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td><strong>Role title<\/strong><\/td>\n<td>Senior Robotics Research Scientist<\/td>\n<\/tr>\n<tr>\n<td><strong>Role purpose<\/strong><\/td>\n<td>Advance robotics autonomy intelligence through rigorous research, scalable evaluation, and transfer of validated methods into integration-ready software components.<\/td>\n<\/tr>\n<tr>\n<td><strong>Top 10 responsibilities<\/strong><\/td>\n<td>1) Define research agenda aligned to product outcomes 2) Design and run reproducible experiments 3) Develop ML models for robotics perception\/policy 4) Build and maintain benchmark suites and regression gates 5) Improve sim2real transfer strategies 6) Perform failure analysis and propose mitigations 7) Collaborate on data strategy and synthetic data 8) Optimize models for deployment constraints 9) Partner with engineering on production transfer 10) Mentor others and lead cross-functional initiatives<\/td>\n<\/tr>\n<tr>\n<td><strong>Top 10 technical skills<\/strong><\/td>\n<td>1) Robotics ML fundamentals 2) Experimentation rigor\/ablation design 3) PyTorch (and\/or JAX) 4) Robotics fundamentals (kinematics\/dynamics\/control) 5) 3D perception &amp; sensor data handling 6) Simulation-driven development 7) RL\/imitation learning (as applicable) 8) Sim2real\/domain adaptation 9) Deployment optimization (ONNX\/TensorRT\/quantization as needed) 10) Benchmarking and automated evaluation design<\/td>\n<\/tr>\n<tr>\n<td><strong>Top 10 soft skills<\/strong><\/td>\n<td>1) Scientific judgment\/integrity 2) Systems thinking 3) Technical communication 4) Cross-functional influence 5) Prioritization under uncertainty 6) Mentorship 7) Resilience\/iterative problem solving 8) Stakeholder empathy 9) Structured decision-making 10) Ownership mentality<\/td>\n<\/tr>\n<tr>\n<td><strong>Top tools\/platforms<\/strong><\/td>\n<td>PyTorch\/JAX; ROS2; Isaac Sim\/Gazebo\/MuJoCo; MLflow or Weights &amp; Biases; Docker; Kubernetes (often); Git + CI (GitHub Actions\/GitLab CI\/Jenkins); ONNX\/TensorRT (deployment dependent); Jira; Confluence\/Notion<\/td>\n<\/tr>\n<tr>\n<td><strong>Top KPIs<\/strong><\/td>\n<td>Benchmark success rate; safety\/constraint violation rate (context-specific); generalization score; regression gate pass rate; experiment cycle time; reproducibility rate; compute efficiency; sim2real transfer delta; runtime latency; stakeholder satisfaction<\/td>\n<\/tr>\n<tr>\n<td><strong>Main deliverables<\/strong><\/td>\n<td>Research roadmap; decision memos; trained model artifacts with provenance; benchmark suite + regression gates; evaluation datasets\/data cards; simulation scenarios; integration-ready modules; failure mode taxonomy; performance and deployment reports; documentation\/model cards; internal playbooks<\/td>\n<\/tr>\n<tr>\n<td><strong>Main goals<\/strong><\/td>\n<td>First 90 days: reproducible baselines + first measurable improvement + integration-ready prototype. 6\u201312 months: major initiative ownership, benchmark automation adoption, production transfer of methods, and IP contributions. Long-term: enable new autonomy capabilities and reduce cost-to-deploy via scalable evaluation and sim2real mastery.<\/td>\n<\/tr>\n<tr>\n<td><strong>Career progression options<\/strong><\/td>\n<td>Staff Robotics Research Scientist; Principal Robotics Research Scientist; Autonomy\/Robotics AI Architect; Robotics Research Manager; Staff Applied Scientist (Embodied AI); Simulation &amp; Synthetic Data Lead; MLOps for Robotics Lead<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n","protected":false},"excerpt":{"rendered":"<p>The **Senior Robotics Research Scientist** advances the company\u2019s robotics intelligence capabilities by inventing, validating, and transferring novel algorithms and learning-based methods into usable software components for real-world or simulated robots. The role blends deep research rigor (hypothesis-driven experimentation, publication-quality evaluation) with engineering pragmatism (reproducible code, measurable performance, integration-ready deliverables).<\/p>\n","protected":false},"author":61,"featured_media":0,"comment_status":"open","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"_joinchat":[],"footnotes":""},"categories":[24452,24506],"tags":[],"class_list":["post-74918","post","type-post","status-publish","format-standard","hentry","category-ai-ml","category-scientist"],"_links":{"self":[{"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/posts\/74918","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/users\/61"}],"replies":[{"embeddable":true,"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/comments?post=74918"}],"version-history":[{"count":0,"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/posts\/74918\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/media?parent=74918"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/categories?post=74918"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.devopsschool.com\/blog\/wp-json\/wp\/v2\/tags?post=74918"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}