* Design, build, and maintain scalable data engineering and data warehousing solutions
* Develop and optimize data pipelines and ETL workflows (e.g., Airflow)
* Work with NoSQL databases, particularly MongoDB
* Implement and maintain Python-based data processing and analytics scripts
* Support Text Mining, Text Data Mining (TDM), and LLM-based applications
* Integrate and manage structured and unstructured scientific/literature-based datasets
* Ensure adherence to FAIR data principles and associated methodologies
* Collaborate closely with interdisciplinary stakeholders, translating technical solutions into clear business value
* Contribute to high-performance computing (HPC) environments where applicable
* Support or contribute to API development (e.g., GraphQL)
Requirements
* Strong experience in Data Engineering, Data Warehousing, and Database design
* Proven experience with NoSQL databases (MongoDB)
* Strong proficiency in Python scripting
* Experience with ETL orchestration tools (e.g., Airflow)
* Solid understanding of Text Analytics, Text Mining, TDM, and LLMs
* Knowledge of FAIR principles
* Familiarity with scientific literature or other large-scale text-based datasets
* Experience working in interdisciplinary teams
* Strong communication skills in English (fluent); German is a plus
* Exposure to high-performance computing environments is an advantage
* Experience with API development (e.g., GraphQL) is a plus
* Biomedical background or education is a plus
#J-18808-Ljbffr