Full Stack Developer Scraping, Noida
Job Description:
Responsibilities
—Capture of massive data on the web and mobile terminals, and the design of architectures such as extraction, deduplication, classification, clustering, and filtering;
—Design and development of distributed web crawlers, able to independently solve various problems encountered in the actual development process;
—Research and development of web page information extraction technology algorithms to improve the efficiency and quality of data capture;
—Analysis and warehousing of crawled data, monitoring of the crawler system and abnormal alarms;
—Designing and developing data collection strategies and anti-shielding rules to improve the efficiency and quality of data collection;
—Design and development of core algorithms according to the system data processing flow and business function requirements;
—Own the development of these tools, services, and workflows to enhance data management, crawl/scrape analysis, reports, and workflows
—Control the testing of the data and the scraping to guarantee compliance, quality, and accuracy
—Monitor the procedure to detect and address any problems with breaks and scale scrapes as necessary
—Create systems for handling large and unstructured data while developing a regulatory update tool for legal clients
—Create a tool that gathers information on regulatory updates for legal clients by using scraping bots on websites, especially regulatory websites
Desired Profile / Criteria / Skills :
Qualifications
—Proficient in Python language with JavaScript/React JS/Next Js, familiar with one or more of the commonly used crawler frameworks, such as Scrapy framework or other Web scraping frameworks, with independent development experience
—Have 2+ years of experience with JavaScript and 1.5+ years of experience working with WebScrape, Crawlers, and Data Extraction.
—Familiar with vertical search crawlers and distributed web crawlers, deeply understanding the principles of web crawlers, having rich experience in data crawling, parsing, cleaning, and storage related projects, and mastering anti-crawler technology and breakthrough solutions.
—Familiar with common data storage and various data processing technologies is preferred
—A solid foundation in data structure and algorithms is preferred
—Experience in distributed crawler architecture design, IP farms and proxy is preferred
—Familiar with commonly used frameworks such as SSH, multithreading, and network communication programming-related knowledge.
—Mentoring experience of 2-3 engineers is preferred.
Key Skills :
Company Profile
PRODUCT BASED- It's a Global student Accommodation marketplace, operating in 265 Countries around the world. Any student traveling abroad for higher studies, --- --- provides them Eco friendly accommodation along with services like International Money Transfer, Guarantor, Forex, Student Bank Account, Room Essentials, Health Insurance, Travel Insurance, Airport Pickup, International SIM, Luggage Storage, Education Loan, Job Search ETC
Apply Now
- Interested candidates are requested to apply for this job.
- Recruiters will evaluate your candidature and will get in touch with you.