Senior Data Cloud Engineer (Full Stack) + Sign on Bonus 5000 Eur
Deutsche Telekom IT Solutions
- Košice, Košický kraj
- Trvalý pracovný pomer
- Plný úväzok
- AIOps Integration:
- Collaborate with AIOps teams to integrate data engineering solutions into AI-driven processes.
- Implement data pipelines that feed into machine learning models for predictive analytics.
- Ensure seamless communication between data infrastructure and AIOps platforms.
- Real-Time Data Processing:
- Develop and maintain real-time data processing solutions using technologies such as Spark Streaming.
- Implement and optimize data streaming processes for immediate insights and actions.
- integrate real-time data into automated decision-making
- Machine Learning Integration:
- Work closely with data scientists to integrate machine learning models into AIOps workflows.
- Implement data features and preprocessing steps necessary for machine learning model input.
- Optimize data processing to support the training and inference phases of machine learning models.
- Data Orchestration, Monitoring and Anomaly Detection:
- Design and implement ETL and ELT processes for efficient data movement and transformation.
- Utilize tools like Prefect, Airflow, and MLFlow for orchestrating data workflows.
- Implement monitoring solutions to track the health and performance of data pipelines.
- Develop and deploy anomaly detection algorithms to identify irregular patterns in data.
- Data Storage and Retrieval:
- Manage and optimize storage solutions for both structured and unstructured data.
- Implement data retrieval interfaces such as AWS Athena for efficient data access.
- ensure data availability and accessibility for analysis
- Cloud Services and Infrastructure:
- Utilize AWS services like EMR, MSK, Kinesis, and others for scalable data processing.
- Optimize cloud infrastructure to support both data engineering and AIOps requirements.
- leverage cloud capabilities for enhanced operational efficiency
- real-time processing (Spark Stream).
- Proficiency in Spark, parallelization (Dask, Pandarallel, Ray).
- Familiarity with message brokers such as RabbitMQ and Kafka.
- Expertise in storage solutions including S3 and Hadoop.
- Hands-on experience with AWS services: AWS EMR, AWS MSK, AWS Kinesis.
- Medium to senior experience in ETL vs. ELT methodologies.
- Expertise in data orchestration tools such as Prefect, Airflow, MLFlow, and AWS Lambda.
- Proficient in AWS services: AWS Glue, AWS Lambda, AWS Step Functions, AWS EventBridge.
- experience with Tabular DBs (SQL, MySQL, PostgreSQL) and Document-based DBs (MongoDB).
- Hands-on experience with AWS services: AWS Athena, AWS DynamoDB, AWS DocumentDB, AWS RDS, AWS Redshift.
- experience with Snowflake, DataBricks.
- Familiarity with visualization tools such as Grafana and ElasticSearch.
- Angular, Python
- Linux env (putty, shell/bash) / VirtualEnv (dockering) / GitLab + source control / Jenkins
- Deploying REST-based web services / APIs
- Web servers such as Nginx, Apache, etc
- ElasticSearch, Kibana, noSQL MongoDB, ArangoDB, Anaconda, datascience libs, MariaDB, PostgreSQL,
- Backend - JavaScript, JSON, NodeJS, Flask, Python, Ruby, Anaconda
- Devops
- AWS, Lambda, S3
- Prefect - data management
- NLP/ML - BERT, DBScan, Numpy, Jupyter Notebook, Hugging Face, GNN, NSP, NER, Facebook Prophet.
- Financial benefits
- Benefits with focus on learning and development
- Benefits with focus on health and sport
- Benefits with focus on family and work - life balance
- Other benefits
- Please be informed that our remote working possibility is only available within Slovakia due to European taxation regulation.