Description
Sybrant Technologies has been in the forefront of transforming its customers into full digital businesses. Though we are small, we grow at a rapid pace due to our capabilities in the contemporary technologies. Sybrant can deep dive in areas such as Mobility, IoT and Analytics in addition to traditional technologies. We can rapidly implement these solutions because of the Products, Frameworks and Partnerships that we have. In addition, our technically sound people and proven processes help in accelerating our customers’ adoption curves. Advantage Sybrant has always been in its nimbleness and delivering high quality yet cost effective solutions. That’s why we are the “Digital Transformation Power” behind our customers. We are a PreludeSys Group Company.
Role
Design, develop, and maintain scalable web data extraction systems and REST APIs to serve clean, structured data from internal databases.
Key Responsibilities
- Develop and maintain web scraping scripts using Python (Requests, BeautifulSoup, Selenium, Scrapy).
- Automate extraction workflows to ensure reliable and repeatable data collection.
- Handle anti-scraping mechanisms such as CAPTCHAs, rotating proxies, headers, and session management.
- Clean, transform, and load extracted data into internal databases.
- Design and build REST APIs to expose processed data from the database.
- Optimize scraping workflows for performance, reliability, and error handling.
- Monitor scraping jobs, troubleshoot failures, and ensure data freshness.
- Maintain documentation for scraping logic, API endpoints, and workflows.
- Collaborate with product and data teams to understand evolving data requirements.
Required Skills
- Strong proficiency in Python.
- Hands-on experience with web scraping tools (Requests, BeautifulSoup, Selenium, Scrapy).
- Good understanding of HTML, DOM structure, XPath, and CSS selectors.
- Experience building REST APIs using FastAPI, Flask, or Django.
- Solid knowledge of SQL and relational databases (MySQL / PostgreSQL).
- Experience handling proxies, cookies, headers, rate limits, and sessions.
- Familiarity with Git and basic CI/CD workflows.
Good to Have
- Understanding of ETL concepts and data engineering workflows.
- Exposure to Airflow or other workflow orchestrators.
- Basic understanding of data pipelines or ML pipelines.