6,762 Data Infrastructure jobs in the United States
Cloud / Data Infrastructure Engineer
Posted 10 days ago
Job Viewed
Job Description
Client : Finance
Location : San Antonio, TX (Onsite 4 days)
Duration : 6+ Month Contract to Hire (must be on w2 without sponsorship now or in the future)
THIS IS A DATA ENGINEER WITH CLOUD/PIPELINE AUTOMATION EXPERIENCE
Need:
Fivetran/Informatica IDMC/ Azure Data Factory/AWS Glue (or similar automation experience)
ETL/ELT
Requirements:
- Extensive Experience with Cloud and Data Platform Infrastructure Engineering.
- Extensive experience with event platforms on-prem or Cloud (Kafka, Redpanda, AWS Kinesis, Azure Event Hubs, GCP Pub/Sub), stream processing frameworks (Flink, Spark Streaming, Amazon Kinesis Data Streams, Google Cloud Dataflow ).
- Experience with ETL/ELT tools like Fivetran, and/or Informatica's cloud product, IDMC; demonstrated ability to design, implement, and optimize data pipelines. This includes experience with and automating connector configuration, data transformation, and schema management.
- Understanding of event-driven architectures, CQRS, and event sourcing.
- Proficiency in monitoring and observability tools (Datadog, Prometheus, Grafana, or similar).
- Strong experience with Infrastructure as Code (IaC) tools (e.g., Terraform, CloudFormation).
- Proven ability to automate tasks and implement CI/CD pipelines.
- Strong problem-solving and communication skills.
- Experience with Kubernetes and container orchestration.
- Experience with various database technologies (SQL and NoSQL).
- Experience in supporting Real-time analytical need.
Required Skills :
Basic Qualification :
Additional Skills :
Background Check : Yes
Drug Screen : Yes
Notes :
Selling points for candidate :
Project Verification Info :MSA: Blanket Approval Received Client Letter: Will Provide"
Exclusive to Apex :Yes
Face to face interview required :No
Candidate must be local :No
Candidate must be authorized to work without sponsorship :Yes
Interview times set : :No
Type of project :
Master Job Title :
Branch Code :
Senior Cloud/Data Infrastructure Engineer
Posted today
Job Viewed
Job Description
Texas, San Antonio 07/01/2025 Contract Active Job Description: Job Summary: We are seeking a Senior Cloud/Data Infrastructure Engineer with 8+ years of experience to lead the design and implementation of scalable data infrastructure and automation in the cloud. The ideal candidate will have hands-on expertise with modern ETL/ELT tools, event-driven architectures, and pipeline automation across cloud platforms such as AWS, Azure, or GCP. Job Responsibilities: Design, develop, and optimize robust cloud-based data platforms Build and maintain real-time and batch data pipelines using tools like Fivetran, Informatica IDMC, Azure Data Factory, or AWS Glue Architect and manage event-driven streaming systems (Kafka, Kinesis, Event Hubs, Pub/Sub) Automate infrastructure provisioning using IaC tools like Terraform or CloudFormation Implement CI/CD pipelines and streamline deployment processes Ensure observability through monitoring tools (Datadog, Prometheus, Grafana) Support real-time analytics by optimizing data ingestion and processing workflows Collaborate with data engineering, DevOps, and application teams for system integration Troubleshoot performance issues and maintain high availability of services Support containerized environments using Kubernetes for orchestration Required Skills: 8+ years of experience in cloud and data platform engineering Proficiency in ETL/ELT tools: Fivetran, Informatica IDMC, Azure Data Factory, or AWS Glue Experience with event streaming platforms: Kafka, Kinesis, Event Hubs, Pub/Sub Strong knowledge of stream processing frameworks: Apache Flink, Spark Streaming, Google Cloud Dataflow, etc. Experience in Infrastructure as Code with Terraform or CloudFormation Familiarity with CI/CD and automation workflows Hands-on experience with Kubernetes and container orchestration Proficiency in monitoring tools like Datadog, Prometheus, Grafana Knowledge of both SQL and NoSQL databases Strong analytical, problem-solving, and communication skills Preferred Skills: Understanding of CQRS, event sourcing, and real-time data processing Experience in supporting real-time analytics use cases Familiarity with multi-cloud environments (AWS, Azure, GCP) Certifications: (Not specified) Education: Bachelor’s degree in computer science, Engineering, or a related field (or equivalent practical experience) #J-18808-Ljbffr
Senior Cloud Data Infrastructure Engineer
Posted today
Job Viewed
Job Description
Join to apply for the Senior Cloud Data Infrastructure Engineer role at ClickHouse
Join to apply for the Senior Cloud Data Infrastructure Engineer role at ClickHouse
About ClickHouse
Established in 2009, ClickHouse leads the industry with its open-source column-oriented database system, driven by the vision of becoming the fastest OLAP database globally. The company empowers users to generate real-time analytical reports through SQL queries, emphasizing speed in managing escalating data volumes. Enterprises globally, including Lyft, Sony, IBM, GitLab, Twilio, HubSpot, and many more, rely on ClickHouse Cloud. It is available through open-source or on AWS, GCP, Azure, and Alibaba.
About ClickHouse
Established in 2009, ClickHouse leads the industry with its open-source column-oriented database system, driven by the vision of becoming the fastest OLAP database globally. The company empowers users to generate real-time analytical reports through SQL queries, emphasizing speed in managing escalating data volumes. Enterprises globally, including Lyft, Sony, IBM, GitLab, Twilio, HubSpot, and many more, rely on ClickHouse Cloud. It is available through open-source or on AWS, GCP, Azure, and Alibaba.
The Cloud AutoScaling team is dedicated to implementing robust vertical and horizontal auto-scaling capabilities within the ClickHouse cloud environment. We seek exceptional software engineers to develop and maintain the auto-scaling infrastructure to transform ClickHouse into a fully functional server-less database solution. Collaborating closely with the core database team, we are actively working on evolving ClickHouse into a cloud-native database system. Additionally, we engage with other cloud teams to drive continuous improvements in cloud infrastructure for enhanced performance and scalability.
What will you do?
- Build a cutting-edge Cloud Native platform on top of the public cloud.
- Improve the metrics pipeline and build algorithms to generate better autoscaling statistics and recommendations.
- Work on the autoscale and Kubernetes operator to support seamless Vertical and Horizontal Auto-scaling.
- Work closely with our ClickHouse core development team and other data plane teams, partnering with them to support auto-scaling use cases as well as other internal infrastructure improvements.
- Architecting and building a robust, scalable, and highly available distributed infrastructure
- 5+ years of relevant software development industry experience building and operating scalable, fault-tolerant, distributed systems.
- Experience building operators with Kubernetes, controller runtime
- Production experience with programming languages like Go, C++
- You are not a stranger to PagerDuty On-call, debugging things in production and are a strong problem-solver
- Expertise with a public cloud provider (AWS, GCP, Azure) and their infrastructure as a service offering (e.g., EC2).
- Experience with Data Storage, Ingestion, and Transformation (Spark, Kafka or similar tools).
- You are passionate about solving data problems at Scale.
- Experience with Python (uv, rye, fastAPI) Data Science (Pandas, NumPy etc) is good to have.
- You have excellent communication skills and the ability to work well within and across engineering teams
$157,000 - $32,000 USD
General US Remote Salary Range
133,450 - 197,200 USD
Los Angeles, CA / Washington, DC - Salary Range
141,300 - 208,800 USD
Seattle, WA - Full Salary Range
141,300 - 208,800 USD
Compensation
For roles based in the United States , you can find above our typical starting salary ranges for this role, depending on your specific location.
The positioning of offers within a certain range depends on various factors, including: candidate experience, qualifications, skills, business requirements and geographical location.
If you have any questions or comments about compensation as a candidate, please get in touch with us at
Perks
- Flexible work environment - ClickHouse is a globally distributed company and remote-friendly. We currently operate in 20 countries.
- Healthcare - Employer contributions towards your healthcare.
- Equity in the company - Every new team member who joins our company receives stock options.
- Time off - Flexible time off in the US, generous entitlement in other countries.
- A 500 Home office setup if youre a remote employee.
- Global Gatherings We believe in the power of in-person connection and offer opportunities to engage with colleagues at company-wide offsites.
As part of our first 500 employees, you will be instrumental in shaping our culture.
Are you interested in finding out more about our culture? Learn more about our values here. Check out our blog posts or follow us on LinkedIn to find out more about whats happening at ClickHouse.
Equal Opportunity & Privacy
ClickHouse provides equal employment opportunities to all employees and applicants and prohibits discrimination and harassment of any type based on factors such as race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws.
Please see here for our Privacy Statement. Seniority level
- Seniority level Mid-Senior level
- Employment type Full-time
- Job function Information Technology
- Industries Technology, Information and Internet, Software Development, and IT Services and IT Consulting
Referrals increase your chances of interviewing at ClickHouse by 2x
Get notified about new Senior Infrastructure Engineer jobs in Austin, Texas Metropolitan Area .
Senior Staff Network Solutions Engineer (REMOTE) Unified Communications Administrator - App PackagerAustin, TX
110,000.00
-
205,000.00
4 days ago
Austin, TX
180,000.00
-
200,000.00
4 months ago
Austin, TX
99,150.00
-
157,400.00
1 month ago
Austin, TX
155,656.00
-
278,361.00
1 month ago
Austin, TX
75,000.00
-
160,000.00
4 days ago
Austin, TX
100,000.00
-
215,000.00
4 days ago
Were unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
#J-18808-LjbffrData Infrastructure Engineer
Posted today
Job Viewed
Job Description
About HeyGen At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences. Learn more at . Visit our Mission and Culture doc here . Position Summary: At HeyGen, we are at the forefront of developing applications powered by our cutting-edge AI research. As a Data Infrastructure Engineer, you will lead the development of fundamental data systems and infrastructure. These systems are essential for powering our innovative applications, including Avatar IV , Photo Avatar , Instant Avatar , Interactive Avatar, and Video Translation. Your role will be crucial in enhancing the efficiency and scalability of these systems, which are vital to HeyGen's success. Key Responsibilities: Design, build, and maintain the data infrastructure and systems needed to support our AI applications. Examples include Multi-modal data processing framework and applications Storage and computation efficiency AI model evaluation and productionization infrastructure Collaborate with data scientists and machine learning engineers to understand their computational and data needs and provide efficient solutions. Stay up-to-date with the latest industry trends in data infrastructure technologies and advocate for best practices and continuous improvement. Assist in budget planning and management of cloud resources and other infrastructure expenses. Qualifications: Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field Proven experience in managing infrastructure for large-scale AI or machine learning projects Excellent problem-solving skills and the ability to work independently or as part of a team. Proficiency in Python Experience optimizing computational workflows Familiarity with AI and machine learning frameworks like TensorFlow or PyTorch. Preferred Qualifications: Experience with GPU computing Experience with distributed data processing system Experience building large scale batch inference system Prior experience in a startup or fast-paced tech environment. What HeyGen Offers Competitive salary and benefits package. Dynamic and inclusive work environment. Opportunities for professional growth and advancement. Collaborative culture that values innovation and creativity. Access to the latest technologies and tools. HeyGen is an Equal Opportunity Employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. Apply for this job * indicates a required field First Name * Last Name * Preferred First Name Email * Phone Resume/CV Enter manually Accepted file types: pdf, doc, docx, txt, rtf Enter manually Accepted file types: pdf, doc, docx, txt, rtf #J-18808-Ljbffr
Data Infrastructure Engineer
Posted today
Job Viewed
Job Description
Job DescriptionJob Description
Worth AI, a leader in the computer software industry, is looking for a talented and experienced Data Infrastructure Engineer to join their innovative team. At Worth AI, we are on a mission to revolutionize decision-making with the power of artificial intelligence while fostering an environment of collaboration, and adaptability, aiming to make a meaningful impact in the tech landscape. Our team values include extreme ownership, one team and creating reaving fans both for our employees and customers.
Worth is looking for a Data Infrastructure Engineer to build and maintain the foundational systems that power our data platform. In this role, you will design, implement, and optimize scalable, reliable, and secure data infrastructure that supports analytics, data science, and product applications across the company.
The ideal candidate is deeply experienced with modern data architectures, cloud platforms, and data orchestration tools. You are passionate about automation, performance tuning, and ensuring high availability of data services. This is a critical role on our data team and offers the opportunity to shape the long-term data strategy of the organization.
Responsibilities
- Design, build, and maintain scalable and resilient data infrastructure in a cloud environment (AWS, Azure, or GCP).
- Develop and maintain ETL/ELT pipelines using orchestration tools such as Airflow, Dagster, or dbt.
- Optimize data workflows for reliability, performance, and cost efficiency across structured and unstructured datasets.
- Manage data lake and data warehouse environments (e.g., Snowflake, BigQuery, Redshift, Delta Lake).
- Ensure data security, privacy, and compliance, including role-based access control, data encryption, and audit logging.
- Collaborate with data scientists, analysts, and product teams to ensure data accessibility, accuracy, and availability.
- Support real-time and batch data processing frameworks, including Kafka, Spark, Flink, or similar tools.
- Monitor, troubleshoot, and improve the observability and performance of data systems using tools like Prometheus, Grafana, or Datadog.
- Maintain CI/CD pipelines for data infrastructure using Terraform, GitHub Actions, or similar tools.
Requirements
- 5+ years of experience in data engineering, infrastructure engineering, or a related field.
- Strong programming skills in Python, Node.js.
- Proficient in SQL and experience with distributed query engines (e.g., Trino, Presto).
- Experience with cloud- data platforms such as AWS Glue
- Hands-on experience with infrastructure-as-code tools (Terraform, Pulumi, CloudFormation).
- Familiarity with containerization and orchestration tools such as Kafka and Kubernetes.
- Solid understanding of data governance, quality frameworks, and data lifecycle management.
- Ability to work in a fast-paced, collaborative environment with a focus on impact and delivery.
- Experience in streaming data architecture and tools like Apache Kafka, Kinesis, or Pub/Sub.
- Background in supporting machine learning or analytics platforms.
- Exposure to data mesh, data contracts, or modern data stack concepts.
- Knowledge of DevOps principles applied to data systems.
Benefits
- Health Care Plan (Medical, Dental & Vision)
- Retirement Plan (401k, IRA)
- Life Insurance
- Unlimited Paid Time Off
- 9 paid Holidays
- Family Leave
- Work From Home
- Free Food & Snacks (Access to Industrious Co-working Membership!)
- Wellness Resources
Data Infrastructure Engineer
Posted 1 day ago
Job Viewed
Job Description
About the Team You'll join the team that's behind OpenAI's data infrastructure that powers critical engineering, product, alignment teams that are core to the work we do at OpenAI. The systems we support include our data warehouse, batch compute infrastructure, streaming infrastructure, data orchestration system, data lake, vector databases, critical integrations, and more. About the Role The Applied Data Platform team designs, builds, and operates the foundational data infrastructure that enables products and teams at OpenAI. You are comfortable with work such as scaling Kubernetes services, OLAP systems, debugging Kafka consumer lag, diagnosing distributed kv store failures, designing a system to retrieve image vectors with low latency. You are well versed with infrastructure tooling such as Terraform, worked with Kubernetes, and have the SRE skill sets. This role is based in San Francisco, CA. We use a hybrid work model of 3 days in the office per week and offer relocation assistance to new employees. In this role, you will: Design, build, and maintain data infrastructure systems such as distributed compute, data orchestration, distributed storage, streaming infrastructure while ensuring scalability, reliability, and security Ensure our data platform can scale reliably to the next several orders of magnitude Accelerate company productivity by empowering your fellow engineers & teammates with excellent data tooling and systems, providing a best in case experience Bring new features and capabilities to the world by partnering with product engineers, trust & safety and other teams to build the technical foundations Like all other teams, we are responsible for the reliability of the systems we build. This includes an on-call rotation to respond to critical incidents as needed You might thrive in this role if you: Have 4+ years in data infrastructure engineering OR Have 4+ years in infrastructure engineering with a strong interest in data Take pride in building and operating scalable, reliable, secure systems Are comfortable with ambiguity and rapid change Have a voracious and intrinsic desire to learn and fill in missing skills-and an equally strong talent for sharing learnings clearly and concisely with others Some of the technologies you'll be working with include Apache Spark, Clickhouse, Python, Terraform, Kafka, Azure EventHub, Vector DBs. About OpenAI OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity. We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or any other legally protected status. OpenAI Affirmative Action and Equal Employment Opportunity Policy Statement For US Based Candidates: Pursuant to the San Francisco Fair Chance Ordinance, we will consider qualified applicants with arrest and conviction records. We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link. OpenAI Global Applicant Privacy Policy At OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology. Compensation Range: $210K - $405K #J-18808-Ljbffr
Data Infrastructure Engineer
Posted 1 day ago
Job Viewed
Job Description
At Jaris, we’re redefining how financial services are delivered by building the modern infrastructure that powers embedded finance for platforms, processors, ISOs, and banks. Our platform streamlines everything from merchant onboarding and underwriting to bank account provisioning, compliance, and money movement — enabling our partners to launch financial products quickly and scale them confidently. By delivering the full stack of enablement tools and value-added financial services, we help our partners unlock new revenue streams and deliver better experiences to their customers. As we expand our impact, we're looking for curious, driven people who want to help modernize the financial ecosystem and support the small businesses that power the economy. About the Role: We are seeking an experienced Data Infrastructure Engineer to join our team and help us continue to build and maintain a robust data platform. The ideal candidate will have a deep understanding of data engineering principles, experience with modern data platforms and tools, and the ability to develop and maintain complex data pipelines and integrations. Responsibilities: Data Platform Work with a cross functional team to identify, rollout and maintain our next generation data platform and warehouse. Work closely with Data Science and Analysts to build pipelines/platform solutions for enabling Analytics, Underwriting, Segmentation, Experimentation and beyond Build and maintain data infrastructure to support Data, ML and AI teams and use cases API Integrations Develop and maintain various pipeline integrations interfacing with Salesforce, Hubspot, Plaid, Datadog and more Ensure seamless data flow and integration between multiple systems and services Data Ingestion Build systems to maintain various pipeline integrations interfacing with Salesforce, Hubspot, Plaid, Datadog and more Develop file parsers and pipelines for ingesting various financial transaction data file formats from providers such as TSYS, Worldpay, and New York Fed SOFR rates Develop frameworks and tools to ensure data accuracy, consistency, and reliability Qualifications: 7+ years of proven experience as Data Engineer in data infrastructure/platform teams Experience in designing, building and maintaining scalable and high availability systems to support end to end data needs with focus on self serviceability Proficiency in SQL and Python Experience working with Data Orchestration/Workflow platforms such as Dagster or Airflow Experience designing and implementing data lakehouse solutions with technologies like AWS, Databricks, Snowflake, Athena Experience with building and optimizing data pipelines for performance and cost using open table file formats like Deltalake, Apache Iceberg Bachelor’s degree in Computer Science, or a related field. Advanced degree preferred. Knowledge of Data Build Tool (DBT), Docker, and AWS ECS Excellent problem-solving skills and attention to detail Strong written and verbal communication skills Nice to Have: Experience building CDC (Change Data Capture) and streaming workflows Experience in financial services or fintech industry Familiarity with financial transaction data and relevant file formats Understanding of data governance and data security best practices Experience working on RESTful web services and delivering API based features Java experience is plus Applicants can expect an annual base compensation in the range of $132,000 to $192,000 USD. This salary range may be inclusive of several career levels at Jaris and will be narrowed during the interview process based on a number of factors, including the candidate’s experience, qualifications, and location. Additional benefits include: Company equity 401(k) plan with a corporate match Employee Assistance Program through Optum Medical, dental, and vision benefits (PPO/HMO/HDHP options) Health & Financial Wellness through a Partnership with Calm, Insperity & MSA (My Secure Advantage) Care Giver Support Program Flexible PTO At Jaris, our corporate culture is based on integrity, respect, collaboration, and transparency, which guide all aspects of our business. As an international company, we combine a diversity of cultures and invest in each individual to provide opportunities for growth and impact. We are an equal opportunity employer and are committed to providing a positive interview experience for every candidate. If accommodations due to a disability or medical condition are needed, connect with us via email at Apply for this job * indicates a required field First Name * Last Name * Email * Phone Resume/CV Enter manually Accepted file types: pdf, doc, docx, txt, rtf Enter manually Accepted file types: pdf, doc, docx, txt, rtf LinkedIn Profile Are you authorized to work in the U.S? * Select. Are you willing to work onsite 3 days per week in Burlingame, CA? * Select. #J-18808-Ljbffr
Be The First To Know
About the latest Data infrastructure Jobs in United States !
Data Infrastructure Engineer
Posted 9 days ago
Job Viewed
Job Description
About HeyGen At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences. Learn more at . Visit our Mission and Culture doc here . Position Summary: At HeyGen, we are at the forefront of developing applications powered by our cutting-edge AI research. As a Data Infrastructure Engineer, you will lead the development of fundamental data systems and infrastructure. These systems are essential for powering our innovative applications, including Avatar IV , Photo Avatar , Instant Avatar , Interactive Avatar, and Video Translation. Your role will be crucial in enhancing the efficiency and scalability of these systems, which are vital to HeyGen's success. Key Responsibilities: Design, build, and maintain the data infrastructure and systems needed to support our AI applications. Examples include Multi-modal data processing framework and applications Storage and computation efficiency AI model evaluation and productionization infrastructure Collaborate with data scientists and machine learning engineers to understand their computational and data needs and provide efficient solutions. Stay up-to-date with the latest industry trends in data infrastructure technologies and advocate for best practices and continuous improvement. Assist in budget planning and management of cloud resources and other infrastructure expenses. Qualifications: Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field Proven experience in managing infrastructure for large-scale AI or machine learning projects Excellent problem-solving skills and the ability to work independently or as part of a team. Proficiency in Python Experience optimizing computational workflows Familiarity with AI and machine learning frameworks like TensorFlow or PyTorch. Preferred Qualifications: Experience with GPU computing Experience with distributed data processing system Experience building large scale batch inference system Prior experience in a startup or fast-paced tech environment. What HeyGen Offers Competitive salary and benefits package. Dynamic and inclusive work environment. Opportunities for professional growth and advancement. Collaborative culture that values innovation and creativity. Access to the latest technologies and tools. HeyGen is an Equal Opportunity Employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. Apply for this job * indicates a required field First Name * Last Name * Preferred First Name Email * Phone Resume/CV Enter manually Accepted file types: pdf, doc, docx, txt, rtf Enter manually Accepted file types: pdf, doc, docx, txt, rtf #J-18808-Ljbffr
Data Infrastructure Engineer
Posted 9 days ago
Job Viewed
Job Description
About HeyGen At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences. Learn more at . Visit our Mission and Culture doc here . Position Summary: At HeyGen, we are at the forefront of developing applications powered by our cutting-edge AI research. As a Data Infrastructure Engineer, you will lead the development of fundamental data systems and infrastructure. These systems are essential for powering our innovative applications, including Avatar IV , Photo Avatar , Instant Avatar , Interactive Avatar, and Video Translation. Your role will be crucial in enhancing the efficiency and scalability of these systems, which are vital to HeyGen's success. Key Responsibilities: Design, build, and maintain the data infrastructure and systems needed to support our AI applications. Examples include Multi-modal data processing framework and applications Storage and computation efficiency AI model evaluation and productionization infrastructure Collaborate with data scientists and machine learning engineers to understand their computational and data needs and provide efficient solutions. Stay up-to-date with the latest industry trends in data infrastructure technologies and advocate for best practices and continuous improvement. Assist in budget planning and management of cloud resources and other infrastructure expenses. Qualifications: Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field Proven experience in managing infrastructure for large-scale AI or machine learning projects Excellent problem-solving skills and the ability to work independently or as part of a team. Proficiency in Python Experience optimizing computational workflows Familiarity with AI and machine learning frameworks like TensorFlow or PyTorch. Preferred Qualifications: Experience with GPU computing Experience with distributed data processing system Experience building large scale batch inference system Prior experience in a startup or fast-paced tech environment. What HeyGen Offers Competitive salary and benefits package. Dynamic and inclusive work environment. Opportunities for professional growth and advancement. Collaborative culture that values innovation and creativity. Access to the latest technologies and tools. HeyGen is an Equal Opportunity Employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. Apply for this job * indicates a required field First Name * Last Name * Preferred First Name Email * Phone Resume/CV Enter manually Accepted file types: pdf, doc, docx, txt, rtf Enter manually Accepted file types: pdf, doc, docx, txt, rtf #J-18808-Ljbffr
Data Infrastructure Engineer
Posted 10 days ago
Job Viewed
Job Description
Worth AI, a leader in the computer software industry, is looking for a talented and experienced Data Infrastructure Engineer to join their innovative team. At Worth AI, we are on a mission to revolutionize decision-making with the power of artificial intelligence while fostering an environment of collaboration, and adaptability, aiming to make a meaningful impact in the tech landscape. Our team values include extreme ownership, one team and creating reaving fans both for our employees and customers.
Worth is looking for a Data Infrastructure Engineer to build and maintain the foundational systems that power our data platform. In this role, you will design, implement, and optimize scalable, reliable, and secure data infrastructure that supports analytics, data science, and product applications across the company.
The ideal candidate is deeply experienced with modern data architectures, cloud platforms, and data orchestration tools. You are passionate about automation, performance tuning, and ensuring high availability of data services. This is a critical role on our data team and offers the opportunity to shape the long-term data strategy of the organization.
Responsibilities
- Design, build, and maintain scalable and resilient data infrastructure in a cloud environment (AWS, Azure, or GCP).
- Develop and maintain ETL/ELT pipelines using orchestration tools such as Airflow, Dagster, or dbt.
- Optimize data workflows for reliability, performance, and cost efficiency across structured and unstructured datasets.
- Manage data lake and data warehouse environments (e.g., Snowflake, BigQuery, Redshift, Delta Lake).
- Ensure data security, privacy, and compliance, including role-based access control, data encryption, and audit logging.
- Collaborate with data scientists, analysts, and product teams to ensure data accessibility, accuracy, and availability.
- Support real-time and batch data processing frameworks, including Kafka, Spark, Flink, or similar tools.
- Monitor, troubleshoot, and improve the observability and performance of data systems using tools like Prometheus, Grafana, or Datadog.
- Maintain CI/CD pipelines for data infrastructure using Terraform, GitHub Actions, or similar tools.
- 5+ years of experience in data engineering, infrastructure engineering, or a related field.
- Strong programming skills in Python, Node.js.
- Proficient in SQL and experience with distributed query engines (e.g., Trino, Presto).
- Experience with cloud-native data platforms such as AWS Glue
- Hands-on experience with infrastructure-as-code tools (Terraform, Pulumi, CloudFormation).
- Familiarity with containerization and orchestration tools such as Kafka and Kubernetes.
- Solid understanding of data governance, quality frameworks, and data lifecycle management.
- Ability to work in a fast-paced, collaborative environment with a focus on impact and delivery.
- Experience in streaming data architecture and tools like Apache Kafka, Kinesis, or Pub/Sub.
- Background in supporting machine learning or analytics platforms.
- Exposure to data mesh, data contracts, or modern data stack concepts.
- Knowledge of DevOps principles applied to data systems.
- Health Care Plan (Medical, Dental & Vision)
- Retirement Plan (401k, IRA)
- Life Insurance
- Unlimited Paid Time Off
- 9 paid Holidays
- Family Leave
- Work From Home
- Free Food & Snacks (Access to Industrious Co-working Membership!)
- Wellness Resources