Join us in revolutionizing the music industry. As the next leader in music technology, Winamp offers a suite of essential services for artists, including the Fanzone, distribution, copyright management, NFTs, and more. For listeners, Winamp provides a unique player experience with direct artist connections, podcasts, radio, and more, all in one place.
We are looking for a passioned Python Data Developer to reinforce our tech team.
What you'll do
1. Data Pipeline Development
- Designing, building, and maintaining data pipelines for efficient data extraction, transformation, and loading (ETL/ELT).
- Integrating data from various sources (e.g., APIs, databases, flat files).
- Writing robust Python scripts for automated data workflows.
2. Database Management
- Creating and managing relational and non-relational databases.
- Writing complex SQL queries for data retrieval and transformation.
- Ensuring data integrity and performance optimization in database systems.
3. Data Analysis & Processing
- Performing data preprocessing tasks such as cleaning, deduplication, and formatting.
- Analyzing large datasets to extract meaningful insights.
- Developing algorithms for data manipulation and aggregation.
4. API Integration and Development
- Consuming and integrating third-party APIs to collect data.
- Developing APIs to expose data for use by other systems or applications.
5. Working with Big Data Technologies
- Implementing scalable solutions using big data frameworks (e.g., Apache Spark, Hadoop).
- Writing Python scripts for distributed data processing.
6. Data Quality and Testing
- Ensuring data quality by creating validation rules and automated testing.
- Debugging and resolving data-related issues in systems and pipelines.
7. Performance Optimization
- Optimizing code and queries for faster data processing
- Ensuring scalability of data solutions for growing datasets
Who you are
- 2+ years of Python experience with the ability to write and maintain object-oriented Python data applications
- Experience in a batch data processing environment or similar with large data jobs using automation
- Experience with ETL pipelines and Data Lakes
- Proficient in SQL and experience with relational databases (e.g., PostgreSQL).
- Experience working with RESTful APIs for data integration
- You are a true data (and music) lover
- Autonomous, able to work independently
- Quick learner that can work and troubleshoot problems independently
- A start-up, agile mindset: you are comfortable with frequent changes and thrive in fast-paced environments
- Fluent in English and French
Nice to have
- Previous experience with PostgreSQL
- Proficiency with Linux, Bash, and understanding of Linux server architecture and operation
- Experience running profilers, debuggers, and general troubleshooting/debugging
- Ability to analyze performance of large and small data queries, jobs, and pipelines to find and fix bottlenecks
- Knowledge of Azure, Docker, and Git version control systems (GitLab and/or GitHub)
- DataOps skills
- Understanding of scikit-learn, TensorFlow, or PyTorch for basic machine learning tasks
- Experience preparing data for AI/ML models