Responsible:
- Develop and maintain scripts for web scraping and data ingestion
- Normalize, clean, and structure large datasets (structured & unstructured)
- Design and maintain scalable databases (NoSQL / SQL)
- Automate workflows for continuous data collection and updates
- Integrate AI tools (e.g., LLMs, code generation assistants) to improve productivity and data processing
- Collaborate with a small, agile, and international team
Requirement:
- Strong programming skills in Python (web scraping, data processing, APIs) (from 4 years of experiences)
- Experience with databases (MongoDB, PostgreSQL, or similar)
- Solid understanding of web scraping frameworks (BeautifulSoup, Scrapy, Playwright, etc.)
- Experience with data cleaning, normalization, and ETL pipelines
- Ability to work with AI-assisted development tools (e.g., GitHub Copilot, ChatGPT, LangChain, LlamaIndex)
- Problem-solving mindset, independence, and willingness to experiment
- Good communication skillsEnglish
Other:
- Experience with maritime datasets or logistics data
- Knowledge of cloud environments (Azure, AWS, GCP)
- Background in data engineering, NLP, or compliance systems
Benefit:
- Remote 100%, long-term contract
- Salary: up to 1400$
- Support from PM and teammate