You will develop the data extraction and processing mechanisms inside our big data infrastructure (Spark, Cassandra), that currently supports the processing of 6GB of data every minute. Your mission will be to improve the overall quality, width and depth of the data collected on companies worldwide.
You will work within our Product Engineering team to define the best technical solution, directly impacting the shape of our products, bringing new ideas to life. You’ll think through all the details of the implementation, as well as different scenarios of product growth and scalability.
You will craft custom ML-based solutions, focused on NLP and text processing, then fit them on TBs of data. You will work with cutting-edge architectures and techniques, then deploy your models in a massively parallel Big Data environment. You will be required to make domain-specific, data-oriented decisions, based on a vast and complete world-view.
You will dedicate the first 30 days to absorb as much as possible and work your way through your first deployment. No matter which path you end up exploring (Big Data, Machine Learning, Software Engineering, Infrastructure or all of them), it puts your work into perspective thanks to our data-driven mindset.
You will develop and implement a comprehensive content strategy with a rich editorial calendar to increase qualified traffic and generate leads.
You will apply first principles thinking and work towards scaling up our business by designing a GTM strategy through performance marketing – articulated end-to-end digital marketing campaigns – consistently identifying new market opportunities to acquire MQLs.
You will apply statistical analysis to find outliers, determine quality KPIs on different samples of data and investigate on how we can improve the overall quality of it. Then communicate and coordinate with the rest of the data team to implement the improvements.