Remote Software Engineer - Data Pipeline Specialist
About the Role
itD Tech is seeking a Remote Software Engineer to design and scale the data pipelines that power next-generation foundation models for machine-generated data, including time series, logs, and large-scale event streams. This Remote Software Engineer role contributes directly to the success of model training and production systems by enabling reliable, high-performance data infrastructure at scale. You'll be part of a dynamic team that is committed to pushing the boundaries of technology and innovation.
What You'll Do
- Build and scale distributed data pipelines for large-scale time series, log data, and high-volume event streams.
- Design and maintain reliable, high-performance Spark and Python workflows to support model training datasets.
- Analyze and resolve performance bottlenecks related to latency, memory utilization, data skew, and throughput.
- Improve data quality, validation processes, and reproducibility for machine learning workloads.
- Partner with machine learning engineers and researchers to enhance data processing capabilities.
Requirements
- 3+ years of experience as a Remote Software Engineer or in a similar role focused on data engineering.
- Strong proficiency in Python and experience with Spark for data processing.
- Deep understanding of distributed systems and data pipeline architecture.
- Experience with performance tuning and optimization of data workflows.
- Ability to work collaboratively in a remote team environment.
Nice to Have
- Familiarity with machine learning frameworks and tools.
- Experience with cloud platforms such as AWS or Azure.
- Knowledge of data governance and compliance standards.
What We Offer
- Comprehensive medical benefits.
- 401(k) plan with company match.
- Paid holidays and flexible vacation policy.
- Remote work with a supportive team culture.
- Opportunities for professional development and growth.
This Remote Software Engineer position at itD Tech offers a unique opportunity to work on cutting-edge data pipelines and machine learning workflows. With competitive salary and comprehensive benefits, it's an attractive role for tech professionals.
Who Will Succeed Here
Proficient in Python and Spark for building and optimizing data pipelines, with hands-on experience in frameworks like Apache Airflow for orchestration.
Strong self-motivation and discipline to excel in a remote work environment, demonstrating effective time management and proactive communication within a distributed team.
Experience with machine learning concepts and distributed systems, coupled with a problem-solving mindset to tackle complex data engineering challenges typical in large-scale event stream processing.
Learning Resources
Career Path
Market Overview
Skills & Requirements
Domain Trends
Industry News
Loading latest industry news...
Finding relevant articles from the last 6 months