7,399 Data Infrastructure jobs in the United States
Data Infrastructure Engineer
Posted 3 days ago
Job Viewed
Job Description
At HeyGen, our mission is to make visual storytelling accessible to all. Over the last decade, visual content has become the preferred method of information creation, consumption, and retention. But the ability to create such content, in particular videos, continues to be costly and challenging to scale. Our ambition is to build technology that equips more people with the power to reach, captivate, and inspire audiences.
Learn more at Visit our Mission and Culture doc here.
Position Summary:
At HeyGen, we are at the forefront of developing applications powered by our cutting-edge AI research. As a Data Infrastructure Engineer, you will lead the development of fundamental data systems and infrastructure. These systems are essential for powering our innovative applications, including Avatar IV, Photo Avatar, Instant Avatar, Interactive Avatar, and Video Translation. Your role will be crucial in enhancing the efficiency and scalability of these systems, which are vital to HeyGen's success.
Key Responsibilities:
- Design, build, and maintain the data infrastructure and systems needed to support our AI applications. Examples include
- Large scale data acquisition
- Multi-modal data processing framework and applications
- Storage and computation efficiency
- AI model evaluation and productionization infrastructure
- Collaborate with data scientists and machine learning engineers to understand their computational and data needs and provide efficient solutions.
- Stay up-to-date with the latest industry trends in data infrastructure technologies and advocate for best practices and continuous improvement.
- Assist in budget planning and management of cloud resources and other infrastructure expenses.
- Bachelor's or Master's degree in Computer Science, Engineering, or a related field
- Proven experience in managing infrastructure for large-scale AI or machine learning projects
- Excellent problem-solving skills and the ability to work independently or as part of a team.
- Proficiency in Python
- Experience optimizing computational workflows
- Familiarity with AI and machine learning frameworks like TensorFlow or PyTorch.
- Experience with GPU computing
- Experience with distributed data processing system
- Experience building large scale batch inference system
- Prior experience in a startup or fast-paced tech environment.
- Competitive salary and benefits package.
- Dynamic and inclusive work environment.
- Opportunities for professional growth and advancement.
- Collaborative culture that values innovation and creativity.
- Access to the latest technologies and tools.
HeyGen is an Equal Opportunity Employer. We celebrate diversity and are committed to creating an inclusive environment for all employees.
Data Infrastructure Engineer
Posted 3 days ago
Job Viewed
Job Description
Headquartered in the Silicon Valley, Meshy is the leading 3D generative AI company on a mission to Unleash 3D Creativity. Meshy makes it effortless for both professional artists and hobbyists to create unique 3D assets-turning text and images into stunning 3D models in just minutes. What once took weeks and $1,000 now takes 2 minutes and $.
Our global team of top experts in computer graphics, AI, and art includes alumni from MIT, Stanford, Berkeley, as well as veterans from Nvidia and Microsoft. With 3 million users (and growing), Meshy is trusted by top developers and backed by premiere venture capital firms like Sequoia and GGV.
- No. 1 popularity, among 3D AI tools, according to A16Z games,
- No. 1 website traffic, among 3D AI tools, according to SimilarWeb (2M monthly visits),
- Leading 3D foundation model, delighted texture & fine geometry,
- 52M funding by Top VCs,
- 2.5M users & 20M models generated!
About the Role:
We are seeking a Data Infrastructure Engineer to join our growing team. In this role, you will design, build, and operate distributed data systems that power large-scale ingestion, processing, and transformation of datasets used for AI model training. These datasets span traditional structured data as well as unstructured assets such as images and 3D models, which often require specialized preprocessing for pretraining and fine-tuning workflows.
This is a versatile role: you'll own end-to-end pipelines (from ingestion to transformation), ensure data quality and scalability, and collaborate closely with ML researchers to prepare diverse datasets for cutting-edge model training. You'll thrive in our fast-paced startup environment, where problem-solving, adaptability, and wearing multiple hats are the norm.
What You'll Do:
- Core Data Pipelines
- Design, implement, and maintain distributed ingestion pipelines for structured and unstructured data (images, 3D/2D assets, binaries).
- Build scalable ETL/ELT workflows to transform, validate, and enrich datasets for AI/ML model training and analytics.
- Pretrain Data Processing
- Support preprocessing of unstructured assets (e.g., images, 3D/2D models, video) for training pipelines, including format conversion, normalization, augmentation, and metadata extraction.
- Implement validation and quality checks to ensure datasets meet ML training requirements.
- Collaborate with ML researchers to quickly adapt pipelines to evolving pretraining and evaluation needs.
- Distributed Systems & Storage
- Architect pipelines across cloud object storage (S3, GCS, Azure Blob), data lakes, and metadata catalogs.
- Optimize large-scale processing with distributed frameworks (Spark, Dask, Ray, Flink, or equivalents).
- Implement partitioning, sharding, caching strategies, and observability (monitoring, logging, alerting) for reliable pipelines.
- Infrastructure & DevOps
- Use infrastructure-as-code (Terraform, Kubernetes, etc.) to manage scalable and reproducible environments.
- Integrate CI/CD best practices for data workflows.
- Data Governance & Collaboration
- Maintain data lineage, reproducibility, and governance for datasets used in AI/ML pipelines.
- Work cross-functionally with ML researchers, graphics/vision engineers, and platform teams.
- Embrace versatility: switch between infrastructure-level challenges and asset/data-level problem solving.
- Contribute to a culture of fast iteration, pragmatic trade-offs, and collaborative ownership.
- Technical Background
- 5+ years of experience in data engineering, distributed systems, or similar.
- Strong programming skills in Python (plus Scala/Java/C++ a plus).
- Solid skills in SQL for analytics, transformations, and warehouse/lakehouse integration.
- Proficiency with distributed frameworks (Spark, Dask, Ray, Flink).
- Familiarity with cloud platforms (AWS/GCP/Azure) and storage systems (S3, Parquet, Delta Lake, etc.).
- Experience with workflow orchestration tools (Airflow, Prefect, Dagster).
- Domain Skills (Preferred)
- Experience handling large-scale unstructured datasets (images, video, binaries, or 3D/2D assets).
- Familiarity with AI/ML training data pipelines, including dataset versioning, augmentation, and sharding.
- Exposure to computer graphics or 3D/2D data processing is strongly preferred.
- Mindset
- Comfortable in a startup environment: versatile, self-directed, pragmatic, and adaptive.
- Strong problem solver who enjoys tackling ambiguous challenges.
- Commitment to building robust, maintainable, and observable systems.
- Kubernetes for distributed workloads and orchestration.
- Data warehouses or lakehouse platforms (Snowflake, BigQuery, Databricks, Redshift).
- Familiarty GPU-accelerated computing and HPC clusters
- Experience with 3D/2D asset processing (geometry transformations, rendering pipelines, texture handling).
- Rendering engines (Blender, Unity, Unreal) for synthetic data generation.
- Open-source contributions in ML infrastructure, distributed systems, or data platforms.
- Familiarity with secure data handling and compliance
- Brain : We value intelligence and the pursuit of knowledge. Our team is composed of some of the brightest minds in the industry.
- Heart : We care deeply about our work, our users, and each other. Empathy and passion drive us forward.
- Gut : We trust our instincts and are not afraid to take bold risks. Innovation requires courage.
- Taste : We have a keen eye for quality and aesthetics. Our products are not just functional but also beautiful.
- Competitive salary, equity, and benefits package.
- Opportunity to work with a talented and passionate team at the forefront of AI and 3D technology.
- Flexible work environment, with options for remote and on-site work.
- Opportunities for fast professional growth and development.
- An inclusive culture that values creativity, innovation, and collaboration.
- Unlimited, flexible time off.
- Competitive salary, benefits and stock options.
- 401(k) plan for employees.
- Comprehensive health, dental, and vision insurance.
- The latest and best office equipment.
Data Infrastructure Engineer
Posted 3 days ago
Job Viewed
Job Description
About the Team
Data Platform at OpenAI owns the foundational data stack powering critical product, research, and analytics workflows. We operate some of the largest Spark compute fleets in production; design, and build data lakes and metadata systems on Iceberg and Delta with a vision toward exabyte-scale architecture; run high throughput streaming platforms on Kafka and Flink; provide orchestration with Airflow; and support ML feature engineering tooling such as Chronon. Our mission is to deliver reliable, secure, and efficient data access at scale and accelerate intelligent, AI assisted data workflows.
Join us to build and operate these core platforms that underpin OpenAI products, research, and analytics.
We're not just scaling infrastructure - we're redefining how people interact with data. Our vision includes intelligent interfaces and AI-assisted workflows that make working with data faster, more reliable, and more intuitive.
About the Role
This role focuses on building and operating data infrastructure that supports massive compute fleets and storage systems, designed for high performance and scalability. You'll help design, build, and operate the next generation of data infrastructure at OpenAI. You will scale and harden big data compute and storage platforms, build and support high-throughput streaming systems, build and operate low latency data ingestions, enable secure and governed data access for ML and analytics, and design for reliability and performance at extreme scale.
You will take full lifecycle ownership: architecture, implementation, production operations, and on-call participation.
You've supported Spark, Kafka, Flink, Airflow, Trino, or Iceberg as platforms. You're well-versed in infrastructure tooling like Terraform, experienced in debugging large-scale distributed systems, and excited about solving data infrastructure problems in the AI space.
This role is based in San Francisco, CA. We use a hybrid work model of 3 days in the office per week and offer relocation assistance to new employees.
In this role, you will:
Design, build, and maintain data infrastructure systems such as distributed compute, data orchestration, distributed storage, streaming infrastructure, machine learning infrastructure while ensuring scalability, reliability, and security
Ensure our data platform can scale by orders of magnitude while remaining reliable and efficient
Accelerate company productivity by empowering your fellow engineers & teammates with excellent data tooling and systems
Collaborate with product, research and analytics teams to build the technical foundations capabilities that unlock new features and experiences
Own the reliability of the systems you build, including participation in an on-call rotation for critical incidents
You might thrive in this role if you:
Have 4+ years in data infrastructure engineering OR
Have 4+ years in infrastructure engineering with a strong interest in data
Take pride in building and operating scalable, reliable, secure systems
Are comfortable with ambiguity and rapid change
Have an intrinsic desire to learn and fill in missing skills, and an equally strong talent for sharing learnings clearly and concisely with others
This role is exclusively based in our San Francisco HQ. We offer relocation assistance to new employees.
About OpenAI
OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.
We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic.
For additional information, please see OpenAI's Affirmative Action and Equal Employment Opportunity Policy Statement.
Qualified applicants with arrest or conviction records will be considered for employment in accordance with applicable law, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer hardware in your possession (including the data contained therein) upon termination of employment or end of assignment; and maintain the confidentiality of proprietary, confidential, and non-public information. In addition, job duties require access to secure and protected information technology systems and related data security obligations.
To notify OpenAI that you believe this job posting is non-compliant, please submit a report through this form. No response will be provided to inquiries unrelated to job posting compliance.
We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link.
OpenAI Global Applicant Privacy Policy
At OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.
Compensation Range: $210K - $405K
Data Infrastructure Engineer
Posted 3 days ago
Job Viewed
Job Description
We all know what a great neighborhood feels like. The sad reality is that every day, our neighborhoods are losing their character, authenticity, and sense of community. From Main Streets to Manhattan, our beloved local businesses are being pushed out at unprecedented rates, replaced with corporate brands or empty storefronts. America is losing its small business backbone - its diversity, energy, and competitive edge.
All while there is a generational opportunity in commercial real estate directly ahead of us. We've been quietly and deliberately building withco to empower millions of small businesses to seize their opportunity to become their own landlords. And if we're successful, we'll have created not only the most financially impactful company of its time, but also neighborhoods that we're proud to live in. Join our mission to make it exciting to be a small business owner in America.
We've raised $30M+ from the world's best investors, including Founders Fund, Canaan, NFX, BoxGroup, and angels like Ken Chenault, Danny Meyer, Venus Williams, Kevin Durant, and more.
Why we need you:
Small businesses have been crushed by landlords and big businesses armed with superior data. That ends now. Your mission? Build the system that levels the playing field - arming America's grittiest entrepreneurs with the data, insights, and strategies to not just compete but dominate.
You'll ensure that the 1P data we collect from customers isn't just organized - it's weaponized to give them the data advantage to win.
What you'll do:
- Build the backbone. Design and develop our data model, pipelines, and practices to support our first digital app.
- Supercharge the experience. Enrich our data model to unlock smarter, more valuable features for customers.
- Own the inputs. Take charge of our data procurement and ingestion process, ensuring our data strategy is reliable and scalable.
- Data obsessed. You have obsessive attention to detail and are maniacal about transforming messy data into something pristine and actionable.
- Technically brilliant. Peers see you as as one of the most technically talented engineers they've worked with.
- Pipeline pro. You've set up seamless 1P and 3P data pipelines before, ideally with financial or geospatial data.
While small business is for everyone, withco's culture is not for everyone. In fact, it's for a very, very, very small group. What makes our culture so special is how much we value accountability and clarity in how we work and curiosity and obsession in how we build. We have a disproportionate number of craftspeople, athletes, ex-founders, PhDs and scientists.
Our Benefits
We're building withco to create generational wealth not only for small business builders, but also our team of builders. We endeavor to make this the most meaningful career experience of your life. In addition, we offer:
- Base salary $50- 220K + meaningful stock options
- Eligibility for carried interest in our real estate vehicles
- 100% company-paid medical, dental, and vision insurance
- A beautiful office in SoHo
- Unlimited PTO and WFH
- Paid parental leave
- 401(k), FSA/HSA, and commuter benefits
- Weekly team lunch and regular team events, always at a small business
We're proud to be an equal opportunity employer. We care about building a uniquely diverse, energetic, inclusive, and service-oriented team - one that understands, cares about, and represents the different small business communities we aim to serve.
Senior Data Infrastructure Engineer
Posted 3 days ago
Job Viewed
Job Description
Gridmatic is a startup trying to help decarbonize the grid by using deep learning to forecast energy prices. We believe better forecasting can have a real-world impact on energy and climate. As extreme weather events get more common, energy prices become increasingly volatile - for instance, in the Texas energy market, prices can go 50x higher than usual in extreme scenarios. When this happens, the ability to forecast these price spikes becomes increasingly important.
We use our machine learning (ML) forecasting and optimization to trade in energy markets, make large-scale battery storage systems more efficient, and sell energy to businesses to protect them from extreme price volatility. Our deep learning models have proven very successful in trading in energy markets, we're successfully operating multiple large batteries (50MW+), and we now sell energy to hundreds of businesses.
We have a very strong team with significant expertise in ML, energy, and optimization. If you're interested in working on complex real-world problems, large-scale data challenges, and applying ML to climate and energy, we'd love to talk to you.
The Role:
We're looking for an engineer to help lead the scaling and reliability of our data infrastructure, which is core to the ML work we do at Gridmatic.
Forecasting energy prices is challenging. We have very effective price forecasting models, but we'd like to go much further - scaling the amount of data we can use in our ML models by a factor of 10-100x by incorporating petabyte-scale weather data, increasing spatial granularity of our price forecasting, and more.
We'd also like someone who can tackle the challenge of scaling and improving reliability of our data platform. We deal with a lot of real-world problems when ingesting data from external sources - downtime, late-arriving data, changing schemas. Improving the reliability of our data pipelines will be critical to our ability to make an impact on the grid.
What we're looking for:
- Experience building the infrastructure for large-scale data processing pipelines (both batch and streaming) using tools like Spark, Kafka, Apache Flink, and Apache Beam.
- Experience designing and implementing large-scale data storage systems (feature store, timeseries DBs) for ML use cases. Strong familiarity with relational databases, data warehouses, object storage, timeseries data, and being adept at DB schema design.
- Experience building data pipelines for external data sources that are observable, debuggable, and verifiably correct. Have dealt with challenges like data versioning, point-in-time correctness, and evolving schemas.
- Strong distributed systems and infrastructure skills. Comfortable scaling and debugging Kubernetes services, writing Terraform, and working with orchestration tools like Flyte, Airflow, or Temporal.
- Strong software engineering skills. Being able to write easy-to-extend and well-tested code.
Our stack includes: Python, GCP, Kubernetes, Terraform, Flyte, React/NextJS, Postgres, BigQuery
- Owning and scaling our data infrastructure by several orders of magnitude. This includes our data pipelines, distributed data processing, and data storage.
- Building a unified feature store for all our ML models.
- Efficient storing and loading hundreds of terabytes of weather data for use in AI-based weather models.
- Processing and storing predictions and evaluation metrics for large-scale forecasting models.
- You have 4+ years of experience building data infrastructure or data platforms
- You have experience with ML infrastructure and have worked at companies that use ML for core business functions
- You're comfortable with ambiguity and a fast-moving environment, and have a bias for action
- You learn and pick up new skills quickly
- You're motivated in making a real-world impact on climate and energy
$180,000 - $50,000 a year
You will also receive Stock Options (ISOs)
Taking care of you today:
- Continuing Education Opportunities
- Flexible PTO
- Medical, Dental and Vision plans with competitive employer contributions
- Pre-Tax commuter benefits
- 1500/year non profit donation matching program through Millie
- Home Office Stipend
Protecting your future for you and your family:
- 401K contribution match up to 4%
- Company-paid parental leave
- Company Paid Life Insurance
- Stock Option Loan Program
FAQ
What's your policy on remote work?
We value the ability to work and collaborate in-person in our early stage as a startup, so Gridmatic has a hybrid policy that will ask you to work in our Cupertino office 3 days a week.
What is your interview process?
You'll usually have a chat with the hiring manager or someone on the team about your background and experience. After that, depending on the role, you'll either have a technical phone screen with an engineer, or work on a take-home project. If that goes well, we'll have you on site in Cupertino for an interview panel with the team, which usually takes about 4 hours.
MTS, Data Infrastructure Engineer
Posted 3 days ago
Job Viewed
Job Description
- Today's Data Scientists are in pain - spending their time manually wrangling data, building models through slow trial and error, taking on painstaking rewrites for deployment, and dealing with countless other frustrating bottlenecks. And the tools they are using for much of this work - e.g. Jupyter notebooks and Pandas - are over a decade old.
- We founded Delphina to change this: our mission is to help the world get better at using data to understand the present and predict the future. Delphina is an AI Agent for Data Science: leveraging a combination of generative AI, large scale optimization, and specialized infrastructure to automate the time-consuming but necessary tasks to build powerful ML models quickly; Delphina will identify relevant data, clean it, train models, and even productionize pipelines.
- Our team has previously led large data science and machine learning teams (covering both applications and infrastructure), built startups, and created successful tools for enterprise ML.
- We're backed by top AI investors, including Fei-Fei Li, Radical VC, and Costanoa VC.
- We're looking for an experienced Data Infrastructure Engineer to join as a Member of our Technical Staffat Delphina.
- As one of our key early hires, you will partner closely with our early team on the direction of our product and drive critical technical decisions. You will have broad impact over the technology, product, and our company's culture.
- You will be responsible for:
- Developing solutions that will orchestrate and support massive quantities of data through stages like ingestion, indexing/mining, transformation, machine learning, online deployment
- Developing platforms that enable scientists, researchers, developers to run ML jobs easily and quickly at scale using the latest technologies
- Defining a consistent continuous integration/deployment model that will encourage cross-functional development teams to self-service application unit testing, deployment and operations
- Influencing and lead cross-functional initiatives that will align the team towards commonly used technologies and methodologies
- Hands on experience in building distributed systems, including real-time streaming and batch data processing
- Proficiency in multiple programming languages relevant for such systems (e.g. Python, Rust, C++, Go, Java)
- Knowledge about what it takes to deploy and operate high availability production systems in the cloud
- Experience designing service-oriented architectures and leveraging various data store technologies
- Energy and ambition to build a product that is surprisingly good in surprising ways
- Intrinsic desire to always be improving our product and yourself. Growth mindset to both stay ahead of the curve and pick up whatever knowledge you're missing to get the job done
- Experience working directly on machine learning - either by partnering with scientists and engineers who are building models or by building models yourself
- Experience leading cross-functional teams through ambiguous problems
- Equity in the company
- Medical, dental, and vision insurance
- 401k
- Unlimited PTO
- Top of the line Apple equipment
- Free lunch in the office
Senior Data Infrastructure Engineer
Posted 3 days ago
Job Viewed
Job Description
IEX (IEX Group, Inc.) is an exchange operator and technology company dedicated to innovating for performance in capital markets. Founded in 2012, IEX launched a new kind of securities exchange in 2016 that combines a transparent business model and unique architecture designed to protect investors. Today, IEX applies its proprietary technology and experience to drive performance across asset classes, serve investors, and advocate for transparent and competitive markets.
At IEX, we are solving complex data problems with best-in-class technology to help achieve our mission of creating fairer financial markets. We are looking for a strong Senior Data Infrastructure Engineer to join our team and help us build new products that will better serve our core customers. This person will have the opportunity to play a key role in developing new platforms for IEX and solve interesting financial data problems.
If you are up for working in a fast-paced environment and a highly collaborative FinTech team elevating the playing field-join us!
About You:
- Detail-oriented
- Analytical mindset with excellent problem-solving abilities
- Excellent communication skills
- Collaborative team player who thrives in a dynamic environment
- Eager to contribute to multiple initiatives in a fast-paced environment
- Design and develop scalable databases and data platforms
- Build GUI tools to help visualize data
- Architect and implement complex data solutions from the ground up
- Maintain and support critical data services, including participating in support rotations
- Perform data maintenance, backup and security checks
- Optimize data access
- Integrate and manage third-party data sources
- Ensure data security in accordance with internal policies and procedures
- 5-10 years of experience in data engineering
- 5+ years of KDB+/Q
- Expertise in Python
- Advanced Linux user
- Preferred:
- Knowledge of databases, development, and internet software
- Knowledge of Java/C++
- Understanding of financial products and equities
- Good to have:
- Experience with KX Dashboards
- Knowledge of working with cloud platforms (AWS, Google and/or Microsoft)
Here at IEX, we are dedicated to an inclusive workplace and culture. We are an Equal Opportunity Employer that does not discriminate on the basis of actual or perceived race, color, creed, religion, alienage or national origin, ancestry, citizenship status, age, disability or handicap, sex, marital status, veteran status, sexual orientation, genetic information or any other characteristic protected by applicable federal, state or local laws. This policy not only complies with all applicable laws and protects workers' rights but is vital to IEX's overall mission and values.
Our job titles may span more than one career level. The starting annual base pay is between $200,000 and $250,000 for this NY-based position. The actual base pay is dependent upon many factors, such as: training, transferable skills, work experience, business needs and market demands. The annual base pay range is subject to change and may be modified in the future. This role is eligible for bonus and equity.
Be The First To Know
About the latest Data infrastructure Jobs in United States !
Data Infrastructure Software Engineer

Posted today
Job Viewed
Job Description
**The Team**
Celonis USA is the fastest-growing division of one of the world's fastest-growing companies.
Come join us on the ground floor on a team led by experienced ex-FAANG coworkers dedicated to developing state-of-the-art technology for Celonis products. The team creates foundational solutions such as a mining engine that enables our customers to interactively execute process mining over enterprise-sized data sets, as well as innovative object-centric modeling tools that empower users to perform process-oriented business analysis in a natural and powerful way. In order to handle the strong growth of our customer base and their demands, we are heavily investing in performance and scalability to ensure that our platform scales efficiently and dynamically in a multi-cloud environment.
Join us if you want to grow in your engineering knowledge, your career and leave your mark on a rapidly growing company.
**The Role**
We are looking for senior backend developers who enjoy building systems with the capability to manage big data. Within our Object-Centric Process Mining team, we are looking for a senior Java software engineer, who develops new features and technologies to support the digital transformation of our core platform to a novel object-centric paradigm.
**The work you'll do**
+ Design and implement critical high-performance systems to handle the ever-growing request rate on our backend infrastructure
+ Process and analyze customer SQLs to generate optimized data pipeline on different backend query engines
+ Evangelize best practices in database usage and end-to-end architecture
+ Set an example for and mentor and grow your teammates
+ Recognize and take action to improve the developer work environment
+ Write clean, understandable, and testable code
+ Provide feedback on the designs and code of peers
+ Manage individual project priorities, deadlines, and deliverables
+ Document development procedures, concepts and knowledge
+ Build, launch, and maintain features in production
+ Help define a fun and inclusive engineering culture
**The qualifications you need**
+ BS in Computer science; Masters or Ph.D. preferred
+ 5+ years of hands-on software engineering experience
+ Experience with SQL processing, query planning and optimization or query engine internals
The base salary range below is for the role in the specified location, based on a Full Time Schedule.
Total compensation package will include base salary + bonus/commission + equity + benefits (health, dental, life, 401k, and paid time off). Please note that the base salary range is a guideline, and that the actual total compensation offer will be determined based on various factors, including, but not limited to, applicant's qualifications, skills, experiences, and location.
The base salary range below is for the role in California, based on a Full Time Schedule.
$150,000-$183,000 USD
**What Celonis Can Offer You:**
+ **Pioneer Innovation:** Work with the leading, award-winning process mining technology, shaping the future of business.
+ **Accelerate Your Growth:** Benefit from clear career paths, internal mobility, a dedicated learning program, and mentorship opportunities.
+ **Receive Exceptional Benefits:** Including generous PTO, hybrid working options, company equity (RSUs), comprehensive benefits, extensive parental leave, dedicated volunteer days, and much more ( . Interns and working students explore your benefits here ( .
+ **Prioritize Your Well-being:** Access to resources such as gym subsidies, counseling, and well-being programs.
+ **Connect and Belong:** Find community and support through dedicated inclusion and belonging programs.
+ **Make Meaningful Impact:** Be part of a company driven by strong values ( that guide everything we do: Live for Customer Value, The Best Team Wins, We Own It, and Earth Is Our Future.
+ **Collaborate Globally:** Join a dynamic, international team of talented individuals.
+ **Empowered Environment:** Contribute your ideas in an open culture with autonomous teams.
**About Us:**
Celonis makes processes work for people, companies and the planet. The Celonis Process Intelligence Platform uses industry-leading process mining and AI technology and augments it with business context to give customers a living digital twin of their business operation. It's system-agnostic and without bias, and provides everyone with a common language for understanding and improving businesses. Celonis enables its customers to continuously realize significant value across the top, bottom, and green line. Celonis is headquartered in Munich, Germany, and New York City, USA, with more than 20 offices worldwide.
Get familiar with the Celonis Process Intelligence Platform by watching this video ( .
**Celonis Inclusion Statement:**
At Celonis, we believe our people make us who we are and that "The Best Team Wins". We know that the best teams are made up of people who bring different perspectives to the table. And when everyone feels included, able to speak up and knows their voice is heard - that's when creativity and innovation happen.
**Your Privacy:**
Any information you submit to Celonis as part of your application will be processed in accordance with Celonis' Accessibility and Candidate Notices ( submitting this application, you confirm that you agree to the storing and processing of your personal data by Celonis as described in our Privacy Notice for the Application and Hiring Process ( .
Please be aware of common job offer scams, impersonators and frauds. Learn more here ( .
Data Infrastructure Software Engineer

Posted 14 days ago
Job Viewed
Job Description
**The Team**
Celonis USA is the fastest-growing division of one of the world's fastest-growing companies.
Come join us on the ground floor on a team led by experienced ex-FAANG coworkers dedicated to developing state-of-the-art technology for Celonis products. The team creates foundational solutions such as a mining engine that enables our customers to interactively execute process mining over enterprise-sized data sets, as well as innovative object-centric modeling tools that empower users to perform process-oriented business analysis in a natural and powerful way. In order to handle the strong growth of our customer base and their demands, we are heavily investing in performance and scalability to ensure that our platform scales efficiently and dynamically in a multi-cloud environment.
Join us if you want to grow in your engineering knowledge, your career and leave your mark on a rapidly growing company.
**The Role**
We are looking for senior backend developers who enjoy building systems with the capability to manage big data. Within our Object-Centric Process Mining team, we are looking for a senior Java software engineer, who develops new features and technologies to support the digital transformation of our core platform to a novel object-centric paradigm.
**The work you'll do**
+ Design and implement critical high-performance systems to handle the ever-growing request rate on our backend infrastructure
+ Process and analyze customer SQLs to generate optimized data pipeline on different backend query engines
+ Evangelize best practices in database usage and end-to-end architecture
+ Set an example for and mentor and grow your teammates
+ Recognize and take action to improve the developer work environment
+ Write clean, understandable, and testable code
+ Provide feedback on the designs and code of peers
+ Manage individual project priorities, deadlines, and deliverables
+ Document development procedures, concepts and knowledge
+ Build, launch, and maintain features in production
+ Help define a fun and inclusive engineering culture
**The qualifications you need**
+ BS in Computer science; Masters or Ph.D. preferred
+ 5+ years of hands-on software engineering experience
+ Experience with SQL processing, query planning and optimization or query engine internals
The base salary range below is for the role in the specified location, based on a Full Time Schedule.
Total compensation package will include base salary + bonus/commission + equity + benefits (health, dental, life, 401k, and paid time off). Please note that the base salary range is a guideline, and that the actual total compensation offer will be determined based on various factors, including, but not limited to, applicant's qualifications, skills, experiences, and location.
The base salary range below is for the role in New York, based on a Full Time Schedule.
$146,000-$178,000 USD
**What Celonis Can Offer You:**
+ **Pioneer Innovation:** Work with the leading, award-winning process mining technology, shaping the future of business.
+ **Accelerate Your Growth:** Benefit from clear career paths, internal mobility, a dedicated learning program, and mentorship opportunities.
+ **Receive Exceptional Benefits:** Including generous PTO, hybrid working options, company equity (RSUs), comprehensive benefits, extensive parental leave, dedicated volunteer days, and much more ( . Interns and working students explore your benefits here ( .
+ **Prioritize Your Well-being:** Access to resources such as gym subsidies, counseling, and well-being programs.
+ **Connect and Belong:** Find community and support through dedicated inclusion and belonging programs.
+ **Make Meaningful Impact:** Be part of a company driven by strong values ( that guide everything we do: Live for Customer Value, The Best Team Wins, We Own It, and Earth Is Our Future.
+ **Collaborate Globally:** Join a dynamic, international team of talented individuals.
+ **Empowered Environment:** Contribute your ideas in an open culture with autonomous teams.
**About Us:**
Celonis makes processes work for people, companies and the planet. The Celonis Process Intelligence Platform uses industry-leading process mining and AI technology and augments it with business context to give customers a living digital twin of their business operation. It's system-agnostic and without bias, and provides everyone with a common language for understanding and improving businesses. Celonis enables its customers to continuously realize significant value across the top, bottom, and green line. Celonis is headquartered in Munich, Germany, and New York City, USA, with more than 20 offices worldwide.
Get familiar with the Celonis Process Intelligence Platform by watching this video ( .
**Celonis Inclusion Statement:**
At Celonis, we believe our people make us who we are and that "The Best Team Wins". We know that the best teams are made up of people who bring different perspectives to the table. And when everyone feels included, able to speak up and knows their voice is heard - that's when creativity and innovation happen.
**Your Privacy:**
Any information you submit to Celonis as part of your application will be processed in accordance with Celonis' Accessibility and Candidate Notices ( submitting this application, you confirm that you agree to the storing and processing of your personal data by Celonis as described in our Privacy Notice for the Application and Hiring Process ( .
Please be aware of common job offer scams, impersonators and frauds. Learn more here ( .
Software Engineer, Data Infrastructure
Posted 3 days ago
Job Viewed
Job Description
Figma is growing our team of passionate creatives and builders on a mission to make design accessible to all. Figma's platform helps teams bring ideas to life-whether you're brainstorming, creating a prototype, translating designs into code, or iterating with AI. From idea to product, Figma empowers teams to streamline workflows, move faster, and work together in real time from anywhere in the world. If you're excited to shape the future of design and collaboration, join us!
The Data Infrastructure team at Figma builds and operates the foundational platforms that power analytics, AI, and data-driven decision-making across the company. We serve a diverse set of stakeholders, including AI Researchers, Machine Learning Engineers, Data Scientists, Product Engineers, and business teams that rely on data for insights and strategy. Our team owns and scales critical data platforms such as the Snowflake data warehouse, ML Datalake, and large-scale data movement and processing applications, managing all data flowing into and out of these platforms.
Despite being a small team, we take on high-scale, high-impact challenges. In the coming years, we're focused on building the foundational infrastructure to support AI-powered products, developing streaming interconnects between our core systems, and revamping our orchestration and financial data architecture with a strong emphasis on data quality, reliability, and efficiency. If you're passionate about building scalable, high-performance data platforms that empower teams across Figma, we'd love to hear from you!
This is a full time role that can be held from one of our US hubs or remotely in the United States.
What you'll do at Figma:- Design and build large-scale distributed data systems that power analytics, AI/ML, and business intelligence.
- Develop batch and streaming solutions to ensure data is reliable, efficient, and scalable across the company.
- Manage data ingestion, movement, and processing through core platforms like Snowflake, our ML Datalake, and real-time streaming systems.
- Improve data reliability, consistency, and performance, ensuring high-quality data for engineering, research, and business stakeholders.
- Collaborate with AI researchers, data scientists, product engineers, and business teams to understand data needs and build scalable solutions.
- Drive technical decisions and best practices for data ingestion, orchestration, processing, and storage.
- 5+ years of Software Engineering experience, specifically in backend or infrastructure engineering.
- Experience designing and building distributed data infrastructure at scale.
- Strong expertise in batch and streaming data processing technologies such as Spark, Flink, Kafka, or Airflow/Dagster.
- A proven track record of impact-driven problem-solving in a fast-paced environment.
- A strong sense of engineering excellence, with a focus on high-quality, reliable, and performant systems.
- Excellent technical communication skills, with experience working across both technical and non-technical counterparts.
- Experience mentoring and supporting engineers, fostering a culture of learning and technical excellence.
- Experience with data governance, access control, and cost optimization strategies for large-scale data platforms.
- Familiarity with our stack, including Golang, Python, SQL, frameworks such as dbt, and technologies like Spark, Kafka, Snowflake, and Dagster.
- Experience designing data infrastructure for AI/ML pipelines.
- The ability to navigate ambiguity, take ownership, and drive projects from inception to execution.
At Figma, one of our values is Grow as you go. We believe in hiring smart, curious people who are excited to learn and develop their skills. If you're excited about this role but your past experience doesn't align perfectly with the points outlined in the job description, we encourage you to apply anyways. You may be just the right candidate for this or other roles.
Pay Transparency Disclosure
If based in Figma's San Francisco or New York hub offices, this role has the annual base salary range stated below.
Job level and actual compensation will be decided based on factors including, but not limited to, individual qualifications objectively assessed during the interview process (including skills and prior relevant experience, potential impact, and scope of role), market demands, and specific work location. The listed range is a guideline, and the range for this role may be modified. For roles that are available to be filled remotely, the pay range is localized according to employee work location by a factor of between 80% and 100% of range. Please discuss your specific work location with your recruiter for more information.
Figma offers equity to employees, as well a competitive package of additional benefits, including health, dental & vision, retirement with company contribution, parental leave & reproductive or family planning support, mental health & wellness benefits, generous PTO, company recharge days, a learning & development stipend, a work from home stipend, and cell phone reimbursement. Figma also offers sales incentive pay for most sales roles and an annual bonus plan for eligible non-sales roles. Figma's compensation and benefits are subject to change and may be modified in the future.
Annual Base Salary Range (SF/NY Hub): $149,000—$350,000 USDAt Figma we celebrate and support our differences. We know employing a team rich in diverse thoughts, experiences, and opinions allows our employees, our product and our community to flourish. Figma is an equal opportunity workplace - we are dedicated to equal employment opportunities regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity/expression, veteran status, or any other characteristic protected by law. We also consider qualified applicants regardless of criminal histories, consistent with legal requirements.
We will work to ensure individuals with disabilities are provided reasonable accommodation to apply for a role, participate in the interview process, perform essential job functions, and receive other benefits and privileges of employment. If you require accommodation, please reach out to These modifications enable an individual with a disability to have an equal opportunity not only to get a job, but successfully perform their job tasks to the same extent as people without disabilities.
Examples of accommodations include but are not limited to:
- Holding interviews in an accessible location
- Enabling closed captioning on video conferencing
- Ensuring all written communication be compatible with screen readers
- Changing the mode or format of interviews
To ensure the integrity of our hiring process and facilitate a more personal connection, we require all candidates keep their cameras on during video interviews. Additionally, if hired you will be required to attend in person onboarding.
By applying for this job, the candidate acknowledges and agrees that any personal data contained in their application or supporting materials will be processed in accordance with Figma's Candidate Privacy Notice.