6,568 Data Training jobs in the United States
Big Data Specialist
Posted today
Job Viewed
Job Description
Big Data Support Engineer – Spark & Kubernetes (L1)
Day 1 Austin, TX Hybride
Full time role
Key Skills:
Apache Spark + Kubernetes L1 Support
Must have very good experience(Need good hands on coders, developers who can are working as L1 support)
PySpark / Python - Must have some experience
Hadoop - Good to have
Detailed JD
• Good with spark and Kubernetes
• Has potential to learn and adapt to new processes
• He was able to provide context on his recent contributions with Python
• Has theoretical understanding of Hadoop, since existing project stack is on S3, but willing to learn.
• Support tasks: recent role involves onboarding and supporting deployment on kubernetes
• Displayed good troubleshooting skills (related to Spark skews, bad nodes on kubernetes)
Big DATA Egineer
Posted today
Job Viewed
Job Description
Job Description:
We are seeking a skilled Data Engineer to design, build, and optimize data pipelines and database solutions supporting critical business operations. The ideal candidate will have strong expertise in Python, SQL, and both time-series and relational databases, along with excellent problem-solving and leadership skills.
Key Responsibilities:
- Design, develop, and maintain robust ETL/data pipelines using Python and SQL.
- Work with time-series databases (InfluxDB, Splunk) and relational databases (PostgreSQL, MySQL).
- Review, redesign, and optimize database schemas and perform data modeling.
- Collaborate with stakeholders to gather requirements and deliver scalable data solutions.
- Monitor pipeline performance, troubleshoot issues, and implement improvements.
- Provide leadership in driving data best practices and strong communication across teams.
- Leverage knowledge of big data tools (Kafka, Spark) and visualization platforms (Power BI).
Qualifications:
- Proven experience in Python, SQL, and database design.
- Strong knowledge of time-series and relational database systems.
- Hands-on experience with schema optimization and data modeling.
- Excellent problem-solving, communication, and leadership skills.
- Experience with Power BI or other data visualization tools (preferred).
- Exposure to big data technologies like Kafka/Spark (nice to have).
Big Data Admin
Posted 1 day ago
Job Viewed
Job Description
The senior data engineer is responsible for orchestrating, deploying, maintaining, and scaling cloud infrastructure targeting big data and platform data management (relational and non-relational).
Big Data Engineer
Posted 3 days ago
Job Viewed
Job Description
Headquartered in the United States, TP-Link Systems Inc. is a global provider of reliable networking devices and smart home products, consistently ranked as the world’s top provider of Wi-Fi devices. The company is committed to delivering innovative products that enhance people’s lives through faster, more reliable connectivity. With a commitment to excellence, TP-Link serves customers in over 170 countries and continues to grow its global footprint.
We believe technology changes the world for the better! At TP-Link Systems Inc, we are committed to crafting dependable, high-performance products to connect users worldwide with the wonders of technology.
Embracing professionalism, innovation, excellence, and simplicity, we aim to assist our clients in achieving remarkable global performance and enable consumers to enjoy a seamless, effortless lifestyle.
KEY RESPONSIBILITIES:
-
Design, develop, and maintain scalable big data infrastructure and pipelines, including data ingestion, cleansing, transformation, and data warehouse modeling for large-scale datasets.
-
Design and maintain vector databases and embedding pipelines to support LLM applications, RAG (Retrieval Augmented Generation) systems, semantic search and agentic capabilities.
-
Collaborate with cross-functional teams to deliver reliable, actionable data solutions that support business and product decisions.
-
Implement and manage batch and streaming ETL/ELT workflows using distributed data processing frameworks Spark and orchestration tools.
-
Participate in data integration and ETL pipeline development, ensuring secure and efficient data processing.
-
Investigate system issues, perform troubleshooting, and assist in optimizing data processing workflows.
Requirements
REQUIRED QUALIFICATIONS
-
Bachelor’s degree in Computer Science, Information Systems, or related field.
-
3-5 years of hands-on experience in data engineering or big data infrastructure, working with large scale datasets in a production environment.
-
Proficiency in Python with experience developing scalable ETL/ELT pipelines or made significant contributions to open source python library
-
Ability to work effectively in a team-oriented environment with good communication and problem-solving skills.
PREFERRED QUALIFICATIONS
- Experience with LLM frameworks and libraries (e.g. LangChain, LlamaIndex) is strongly preferred
Benefits
Salary Range: $100,000 - 150,000
-
Free snacks and drinks, and provided lunch on Fridays
-
Fully paid medical, dental, and vision insurance (partial coverage for dependents)
-
Contributions to 401k funds
-
Bi-annual reviews, and annual pay increases
-
Health and wellness benefits, including free gym membership
-
Quarterly team-building events
At TP-Link Systems Inc., we are continually searching for ambitious individuals who are passionate about their work. We believe that diversity fuels innovation, collaboration, and drives our entrepreneurial spirit. As a global company, we highly value diverse perspectives and are committed to cultivating an environment where all voices are heard, respected, and valued. We are dedicated to providing equal employment opportunities to all employees and applicants, and we prohibit discrimination and harassment of any kind based on race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local laws. Beyond compliance, we strive to create a supportive and growth-oriented workplace for everyone. If you share our passion and connection to this mission, we welcome you to apply and join us in building a vibrant and inclusive team at TP-Link Systems Inc.
Please, no third-party agency inquiries, and we are unable to offer visa sponsorships at this time.
Big Data Developer
Posted 3 days ago
Job Viewed
Job Description
Title-Big Data Developer
Location: jersey city, NJ
Contract: 6-18 months
Interview process: screening with apex then one and done interview with client
Must have: Core java, Hadoop, scala, oracle, devops
Nice to have: apache spark
Project: get data from multiple teams. Then they build and consolidate it, data can be used by trading desk with the bank, trading desk from the data and put it into the system, one more level into data because as the system grows, its all getting changed, legacy platform, hard to do this within oracle and that's the reason they are changing it,
Technology stack: put data into Hadoop platform, basic understanding of Hadoop, apache spark. Scala, big data java, hibernate, bundles.
Job title: big data developer
Must haves:
-Core java
-apache spark
-hadoop
-any mapping tool (how the system goes between core and database)
-Scala
-Oracle
-devops (can devops tool is fine, they use Jenkins)
Interview process: screening then one and done video,
Big Data Developer
Posted 3 days ago
Job Viewed
Job Description
Big Data Developer
Jersey City- NJ
Long Term Contract
Job Description:
- Hadoop, Java/Scala, Spark" Big Data Developer - (Count : 1)
- 10 Plus Years Overall IT experience, Big Data Developer with experience in building Java/Scala, Spark based application
- At least 5 Years of strong Development experience along with troubleshooting skills with technologies like Spark, Java/Scala and Hive.
- Experience with Hadoop stack, HDFS, Yarn resourcing, Zookeeper interactions are good to have.
- Good SQL knowledge on Hive/Impala
- Basic Data warehousing concepts like SCD2 will be an added advantage.
- Experience with Kafka, Elastic will be a big plus
- Ability to work under pressure and manage deadlines or unexpected changes in expectations or requirement
Big Data Developer
Posted 3 days ago
Job Viewed
Job Description
- Develop and maintain data pipelines using Java, PySpark, and SQL
- Design and optimize data architectures and data warehousing solutions
- Hands-on experience with Big Data technologies: Spark, Hive, Hadoop
- Utilize GCP managed services, especially BigQuery and DataFlow
- Implement version control using Git and support CI/CD pipelines
- Strong problem-solving skills with a focus on scalable, efficient data processing
Be The First To Know
About the latest Data training Jobs in United States !
Big Data Analyst
Posted 3 days ago
Job Viewed
Job Description
Position Type: Contract
Hadoop/Dataproc Platform Analyst
We are seeking a highly skilled Hadoop/Dataproc Expert to support a strategic initiative focused on the reverse engineering, analysis, and retirement/migration of our existing Hadoop/Google Cloud Dataproc platform. This role is critical to enabling the project team to fully understand the current data ecosystem and prepare for a seamless transition.
Key Responsibilities:
Platform Analysis & Assessment: Conduct a comprehensive review of the Hadoop/Dataproc environment, including cluster configurations, resource usage, and job execution patterns. Identify and document all active components, services, and dependencies.
Data & Metadata Extraction: Pull and catalog all data sources, datasets, and associated metadata across the platform. Map data lineage and relationships to support reverse engineering efforts.
Feed & Log Analysis: Analyze incoming and outgoing data feeds, including batch and streaming pipelines. Review system logs, audit trails, and job histories to identify integration points and data flow patterns.
Collaboration & Documentation: Work closely with data engineers, architects, and project leads to provide insights and technical guidance. Document findings in a structured format to support migration planning and platform decommissioning.
Required Skills & Experience:
- Proven expertise in Hadoop ecosystem (HDFS, Hive, Pig, Spark, etc.) and Google Cloud Dataproc.
- Strong understanding of platform management, including cluster operations, job orchestration, and performance tuning.
- Familiarity with cloud migration strategies and legacy platform retirement best practices.
- Excellent analytical, communication, and documentation skills.
#LI-JU1
Big Data Developer
Posted 3 days ago
Job Viewed
Job Description
Duration-Contract
Location- Richardson, TX~ Woonsocket, RI
JD
- Strong in SQL and Python, 5+ years' experience.
- Experience with big data frameworks (i.e. Hadoop and Spark), 5+ years' experience. Experience building automated data pipelines, 5+ years of experience.
- Experience performing data analysis and data exploration, 5+ years' experience.
- Experience working in an agile delivery environment.
- Strong critical thinking, communication, and problem solving skills.
- Experience with Google Cloud Platform.
- Experience with GCP services (GCS Bucket, Cloud Functions, Dataproc, Dataflows, Pub-Sub)Experience working in multi-developer environment, using version control (i.e. Git)
- Experience with orchestrating pipelines using tools (i.e. Airflow, Azure Data Factory)
- Experience with real-time and streaming technology (i.e. Azure Event Hubs, Azure Functions Kafka, Spark Streaming)
- Experience with API build. Exposure/understanding DevOps best practice and CICD (i.e. Jenkins) Exposure/understanding of containerization (i.e. Kubernetes, Docker)
- Digital : Python, Snowflake
Pallavi Verma
Sr. Technical Recruiter | Syntricate Technologies Inc.
Direct : |
Email : | Web:
We're hiring! connect with us on LinkedIn nd visit our Jobs Portal
Minority Business Enterprise (MBE) Certified | E-Verified Corporation | Equal Employment Opportunity (EEO) Employer
This e-mail message may contain confidential or legally privileged information and is intended only for the use of the intended recipient(s). Any unauthorized disclosure, dissemination, distribution, copying or the taking of any action in reliance on the information herein is prohibited. Please notify the sender immediately by email if you have received this email by mistake and delete this e-mail from your system. You have received this email as we have your email address shared by you or from one of our data sources or from our member(s) or subscriber(s) list. If you do not want to receive any further emails or updates, please reply and request to unsubscribe .
Big Data Developer
Posted 3 days ago
Job Viewed
Job Description
Location: Jersey City, NJ/Brokerage, NJ
Type : Contract
Job Description:
- Strong expertise in Big Data technologies, including hands-on experience with Hadoop ecosystem components like Sqoop and Spark for large-scale data processing and management?
- Proficient in Java programming, with experience in Spring, Spring Boot, and REST services to develop robust backend solutions and integrate with distributed systems?
- Skilled in Angular framework for building dynamic and responsive front-end applications that enhance user experience?
- Experience with Oracle database, including designing, querying, and optimizing database performance?
- Knowledge of event-driven architecture is an added advantage?
- Ability to collaborate effectively across teams and contribute beyond project tasks by participating in mentorship, conducting interviews, and supporting firmwide initiatives aligned with BNY's culture of ownership and continuous learning?
- Strong financial experience with brokerage, custody services
Diverse Lynx LLC is an Equal Employment Opportunity employer. All qualified applicants will receive due consideration for employment without any discrimination. All applicants will be evaluated solely on the basis of their ability, competence and their proven capability to perform the functions outlined in the corresponding role. We promote and support a diverse workforce across all levels in the company.