Information Technology
Full-Time
eMedEvents Global Marketplace for CMECE
Overview
Python Developer Web Scraping & Data Processing
Experience : 3+ Years
Employment Type : Full-time
Job Overview
We are seeking a skilled and detail-oriented Python Developer with 3+ years of hands-on experience in web scraping, document parsing (PDF, HTML, XML), and structured data extraction. You will be a vital part of a core team focused on aggregating biomedical content from diverse sources, including grant repositories, scientific journals, conference abstracts, treatment guidelines, and clinical trial databases. This role demands strong technical proficiency in various parsing and scraping libraries, along with solid data processing and integration skills.
Key Responsibilities
Experience : 3+ Years
Employment Type : Full-time
Job Overview
We are seeking a skilled and detail-oriented Python Developer with 3+ years of hands-on experience in web scraping, document parsing (PDF, HTML, XML), and structured data extraction. You will be a vital part of a core team focused on aggregating biomedical content from diverse sources, including grant repositories, scientific journals, conference abstracts, treatment guidelines, and clinical trial databases. This role demands strong technical proficiency in various parsing and scraping libraries, along with solid data processing and integration skills.
Key Responsibilities
- Develop scalable Python scripts to effectively scrape and parse biomedical data from a wide range of web sources, including websites, pre-print servers, citation indexes, scientific journals, and treatment guidelines.
- Build robust modules specifically for splitting multi-record documents (such as PDFs, HTML, and other formats) into individual, manageable content units.
- Implement NLP-based field extraction pipelines utilizing libraries like spaCy, NLTK, or advanced regex for precise metadata tagging.
- Design and automate complex data acquisition workflows using schedulers and orchestrators like cron, Celery, or Apache Airflow for periodic scraping and content updates.
- Store parsed and processed data efficiently in both relational (PostgreSQL) and NoSQL (MongoDB) databases, ensuring optimal schema design for performance and scalability.
- Ensure robust logging, comprehensive exception handling, and rigorous content quality validation across all data processing and scraping workflows.
- 3+ years of hands-on experience in Python, particularly focused on data extraction, transformation, and loading (ETL).
- Strong command over web scraping libraries, including :
- BeautifulSoup
- Scrapy
- Selenium
- Playwright
- Proficiency in PDF parsing libraries, such as :
- PyMuPDF
- pdfminer.six
- PDFPlumber
- Experience with HTML/XML parsers: lxml, XPath, html5lib.
- Familiarity with regular expressions, NLP concepts, and advanced field extraction techniques.
- Working knowledge of SQL and/or NoSQL databases (MySQL, PostgreSQL, MongoDB).
- Understanding of API integration (RESTful APIs) for interacting with structured data sources.
- Experience with task schedulers and workflow orchestrators (cron, Apache Airflow, Celery).
- Proficiency in version control using Git/GitHub and comfort working in collaborative development environments.
- Exposure to biomedical or healthcare data parsing (scientific abstracts, clinical trials data, drug labels).
- Familiarity with cloud environments like AWS (specifically Lambda, S3 for data storage and processing).
- Experience with data validation frameworks and building robust QA rules for data quality.
- Understanding of ontologies and taxonomies (UMLS, MeSH) for structured content tagging.
Similar Jobs
View All
Talk to us
Feel free to call, email, or hit us up on our social media accounts.
Email
info@antaltechjobs.in