At Vention, we assemble senior-level, dedicated teams of developers to help fast-growing startups and innovative enterprises drive impact and achieve their goals. We’ve delivered solutions across multiple domains, including FinTech, PropTech, AdTech, HealthTech, e-commerce, and more.
Our Data team works with clients to create data platforms from scratch or modify and update existing platforms. The tech stack depends on the project, but we mainly use Spark (along with Scala, Python, or Java) – as well as Apache Kafka, Apache Cassandra, Apache Hadoop, Apache Parquet, and AWS.
Internal knowledge transfer activities are conducted within the Data Engineering Family (which includes data practice & data competency) – it is a space for all of our specialists to share their experiences, learn new skills, host meetups, mentor others, and more.
What will you be responsible for?
- Creation of a data processing pipeline to automate the ML model's education based on constantly refreshing historical data from different sources.
- Migration from SQL/NoSQL databases to Cloud Data Warehouses (Snowflake/AWS Redshift/ Google Big Query)
- Manage various databases and large-scale processing systems in order to provide easy access to the required information
- Design, create, and maintain analytical infrastructure
What skills should you have?
- At least 3+ years of professional experience
- Experience with Data Engineering & Spark
- Solid understanding of data modeling and data storage
- Solid understanding of SQL and NoSQL; and one of the following languages: Python, Java, Scala, R, Bash, C++, .NET
- Hands-on experience with AWS, GCP, or Azure
- Hands-on experience with building ETL processes
- Experience working with Docker, Kubernetes
- A B1 level of English or higher
Will be a plus:
- Apache AirFlow experience