Job Description
Job Role : Big Data Engineer
Job Location : Dallas, TX (Hybrid)
Visa: H1B
Contract: W2, H1T
EXP: 12+
Design, develop, and maintain large-scale data systems and pipelines to process and analyze massive amounts of structured and unstructured data.
Implement efficient, fault-tolerant, and scalable data architectures using big data technologies such as Hadoop, Spark, Kafka, and others.
Work closely with data scientists, analysts, and software engineers to understand data requirements and deliver solutions that support data-driven decision-making.
Build and maintain data integrations and ETL pipelines to collect, clean, and transform data from various sources.
Optimize data processing workflows to ensure high performance and low-latency data pipelines.
Collaborate with cross-functional teams to define data strategies, architecture, and frameworks that support the company s data-driven vision.
Troubleshoot, debug, and resolve technical issues related to data workflows and infrastructure.
Stay up-to-date with emerging trends and technologies in big data and cloud computing to enhance system capabilities.
Required Skills and Qualifications:
Bachelor s degree in Computer Science, Engineering, or a related field (Master’s preferred).
5+ years of experience in big data engineering, data architecture, or related fields.
Proficiency in big data technologies (Hadoop, Spark, Kafka, etc.).
Experience with programming languages such as Java, Python, or Scala.
Strong experience with SQL and NoSQL databases (e.g., MySQL, MongoDB, Cassandra).
Familiarity with cloud platforms (AWS, Google Cloud, Azure) and tools for data processing and storage (e.g., S3, Redshift, Big Query).
Proven experience in designing and maintaining scalable and high-performance data pipelines.
Knowledge of data warehousing concepts and data modeling.
Strong problem-solving and debugging skills, with the ability to work under pressure to deliver results.
Excellent communication skills and the ability to work collaboratively in a team-oriented environment.
Preferred Qualifications:
Experience with data streaming technologies (e.g., Apache Flink, Apache Beam).
Experience with containerization technologies (Docker, Kubernetes).
Familiarity with machine learning and AI frameworks.
Knowledge of DevOps practices and continuous integration/continuous deployment (CI/CD).