Data Systems Engineer – ELK/Kafka/Linux
Listed on 2026-02-24
-
Software Development
Data Engineer
Dice is the leading career destination for tech experts at every stage of their careers. Our client, MARKS IT SOLUTIONS LLC, is seeking the following. Apply via Dice today!
This is a hands-on, client-facing engineering role within the Real-Time Operations Intelligence (RTOI) team, supporting large-scale streaming data platforms. The position requires close collaboration with cross-functional teams and support for hundreds of internal users, so strong communication skills and a team-oriented mindset are essential.
Below is a brief overview of the position:
Job Title: Data Systems Engineer – ELK/Kafka/Linux
Location: Alpharetta, GA or Menlo Park, CA (Onsite)
Experience Level: 7–15 years
Required Qualifications- 7+ years of overall IT/application development experience
- 5+ years of hands-on coding experience in at least one language:
Python (preferred), Ruby, Shell, Java, C/C++, or Go - Minimum 2+ years of data engineering experience with Kafka
- Strong working knowledge of Linux platform (application deployment, debugging, performance tuning, CPU/memory troubleshooting)
- Experience building and supporting real-time ETL/streaming pipelines using Kafka and ELK (Elasticsearch/Logstash/Kibana)
- Experience running, deploying, and supporting applications in large-scale Linux cluster environments
- Strong understanding of distributed systems architecture, scalability, and Kafka ecosystem design trade-offs
- SQL and database experience
- Ability to handle full software development lifecycle (requirements, design, implementation, testing, deployment, and support)
- Strong debugging and production support experience (ensuring jobs are up and running for hundreds of users)
- Excellent communication skills, team-oriented mindset, and strong curiosity/learning ability
- Experience with Snowflake database
- Spark data processing experience
- Hadoop ecosystem exposure
- Observability and data analysis background
- AWS or other cloud technologies
- ELK certification
- Financial industry experience (nice to have)
- Bachelor’s degree (strong plus, not required)
- Design and develop large-scale streaming ETL pipelines using Kafka and ELK
- Deploy and manage applications in Linux-based production environments
- Ensure scalability and reliability within Kafka cluster environments
- Support hundreds of internal stakeholders and dashboards with real-time data needs
- Troubleshoot and debug production issues across distributed systems
- Work across on-prem and cloud-based platforms
- Contribute to development, deployment, and ongoing operational support
This role requires a well-rounded data engineer with strong core application development skills—not just scripting ability—who thrives in large-scale distributed Linux environments.
#J-18808-Ljbffr(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).