Data Engineer

Core team
$80K - $150K/yr compensation

Required Skills

Python
Pandas
SQL
AI/ML
About micro1
micro1 connects domain experts to the development of frontier AI models. Real-world expertise is turned into training data, evaluations, and feedback loops that improve how models perform. AI labs and enterprises use micro1 to train models and build reliable AI agents through advanced evaluations and reinforcement learning environments. Experts contribute directly to how AI systems learn, reason, and perform across domains like finance, healthcare, engineering, and more. Our platform identifies and vets top talent through an AI recruiter, enabling high-quality contributions at scale.
Our goal is to enable 1 billion people to do meaningful work by applying their expertise to AI. We’ve raised $40M+ in funding, and our AI recruiter has powered over 1 million AI-led interviews as our global network of experts grows into the human intelligence layer for AI.

Job Description

Job Title: Data Engineer


Job Type: Full-time


Location: Remote


The Role

We are looking for a Data Engineer to support data infrastructure and experimentation in an AI research environment. In this role, you will build reliable data pipelines, explore datasets, and help transform raw data into structured formats that enable research and model development.


Key Responsibilities

  1. Design, build, and maintain scalable data pipelines to ingest, process, and transform data from multiple sources.
  2. Collaborate with AI researchers and data scientists to structure and prepare datasets for experimentation and model training.
  3. Develop and maintain data models, schemas, and storage systems optimized for large-scale datasets.
  4. Write efficient SQL queries and Python scripts to extract, transform, and analyze data.
  5. Ensure data quality, integrity, and reliability across data pipelines and storage layers.
  6. Implement data validation, monitoring, and automation workflows that support iterative research cycles.


Required Skills and Qualifications

  1. Strong proficiency in Python and SQL.
  2. Experience designing and maintaining ETL / ELT pipelines.
  3. Solid experience with data manipulation libraries such as Pandas and NumPy.
  4. Experience working with structured and semi-structured datasets.
  5. Familiarity with relational databases such as PostgreSQL or MySQL.
  6. Strong analytical thinking and ability to work in collaborative research-driven environments.
  7. Excellent written and verbal communication skills.


Nice to Have

  1. Exposure to AI/ML workflows or research environments.
  2. Experience with data visualization tools such as Matplotlib, Seaborn, or Plotly.
  3. Familiarity with LLM-related data workflows (datasets for training, evaluation, or prompt experimentation).

Apply now

Please note that after completing the interview process, you’ll be added to our talent pool and considered for this and other roles that match your skills.

Have any questions? See FAQs

Refer and Earn$1500