Responsible for the building, deployment, and maintenance of mission-critical analytics solutions that process data quickly at big data scales
Responsible for design and implementation data integration pipelines on AWS Big Data tech stack, using Apache Spark, Hive, HBase, ELK, PostgreSQL, Lambda.
Contributes design, code, configurations, and documentation for components that manage data ingestion, real-time streaming, batch processing, data extraction, transformation, and loading across multiple data storages.
Owns one or more key components of the infrastructure and works to continually improve it, identifying gaps and improving the platform’s quality, robustness, maintainability, and speed.
Cross-trains other team members on technologies being developed, while also continuously learning new technologies from other team members.
Interacts with engineering teams and ensures that solutions meet customer requirements in terms of functionality, performance, availability, scalability, and reliability.
Performs development, QA, and dev-ops roles as needed to ensure a total end to end responsibility of solutions.
Contribute in CoE activities and community building, participate in conferences, provide excellence in exercise and best practices.
Requirements (Hard skills requirements):
3+ years of experience coding in SQL, Python, PySpark, Scala, with solid CS fundamentals including data structure and algorithm design
2+ years contributing to production deployments of large backend data processing and analysis systems as a team lead
1+ years of experience in cloud data platforms (AWS)
Knowledge of professional software engineering best practices for the full software
Knowledge of Data Warehousing, design, implementation, and optimization
Would be a plus:
1+ years of hands-on implementation experience working with a combination of the following technologies: Hadoop, Hive, Spark, SQL, and NoSQL data-warehouses such as Hbase.
Knowledge of Data Quality testing, automation, and results visualization
Knowledge of BI reports and dashboards design and implementation
Knowledge of development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations.
Experience participating in an Agile software development team, e.g. SCRUM
Experience designing, documenting, and defending designs for key components in large distributed computing systems
A consistent track record of delivering exceptionally high-quality software on large, complex, cross-functional projects
Demonstrated ability to learn new technologies quickly and independently
Ability to handle multiple competing priorities in a fast-paced environment
Undergraduate degree in Computer Science or Engineering from a top CS program required. Masters preferred.
Experience with supporting data scientists and complex statistical use cases highly desirable
Desirable (Hard skills):
Understanding of cloud infrastructure design and implementation
Experience in data science and machine learning
Experience in backend development and deployment
Experience in CI/CD configuration
Good knowledge of data analysis in enterprises
Personal skills (soft skills):
A curious mind and willingness to work with the client in a consultative manner to find areas to improve
Upper-Intermediate or Advanced English
Good analytical skills
Good team player motivated to develop and solve complex tasks
Self-motivated, self-disciplined and result-oriented
Strong attention to details and accuracy
Full-time workload on remote basis
Long-term assignment (minimum 6 month contract)
Project starts in Feb/Mar 2021 (upon your availability)
Worthy reward based on your skills and experience
Contact me to get more details at Skype: live:mi_1582 or firstname.lastname@example.org
If this sounds like you, we'd welcome your application!