Data scrapingJobs
I want to replace several manual reporting routines with an end-to-end AI workflow that ingests data from our internal finance databases and live web sources, then produces clear, timely analytics for management. Reporting and analytics are the sole focus—no transaction execution—so the system must excel at pulling, cleaning, and interpreting numbers rather than booking them. We also want to compare legal documents vs term sheets and excel spreadsheets Data sources • Company databases (SQL, flat files, Excel exports) - Dropbox all our files are in drop box • Extensive web scraping for competitor benchmarks and investment-market signals If you have ideas for safely adding external financial APIs later, let me know, but the two feeds abov...
...leaves the shop in “as-new” condition. What matters most to me is proven experience integrating modern CNC controls into legacy iron without compromising accuracy or rigidity. If you have recently worked with Siemens, Fanuc, Heidenhain—or another controller you feel is better suited—let’s discuss the architecture you would recommend and why. Likewise, I’d like to hear how you handle mechanical scraping, bearing replacement, ball-screw refurbishment, and alignment so we can agree on tolerances before any metal is cut. Deliverables I expect: • Detailed retrofit plan and timeline, including control and electrical schematics • Procurement list with lead times for every replacement part • Mechanical overhaul report (inspections,...
...high-performance, real-time flight data aggregation API and the complete engineering setup of the physical server infrastructure. Never get blocked. Bypass any captcha or security meassures they have. Scope of Work: Scraping Microservice: Building the Python-based API to fetch live flight data using advanced anti-bot evasion (TLS Fingerprinting & Headless Browsers). Infrastructure Engineering: Remote configuration of the Bare Metal Dell Server (Dual Xeon) and the Mobile Proxy Farm (OptiPlex + Androids). Integration: Setting up the POST /search endpoints for the backend team. Milestone Breakdown ($2,250 Total): Milestone 1: Sky scanner Scraper Engine ($800) Development of the Python API for Skyscanner. Implementation of curl_cffi / TLS fingerprinting for l...
...Automatically find and extract text from Privacy Policy and Terms & Conditions pages. AI Analysis: Send the extracted data to OpenAI API (using a custom expert-level prompt) to identify legal gaps or outdated clauses. PDF Generator: Create a branded, professional audit report (e.g., using ReportLab or WeasyPrint). 3. Infrastructure: The developer will have Full Root Access to a VPS (Ubuntu). Setup of the Python environment (FastAPI or Flask) as a private API. Database: Connection between WordPress (MySQL) and the Python scanner to store results. Technical Stack: Frontend: WordPress (Elementor/Astra), PHP, JavaScript. Backend: Python 3.x, FastAPI/Flask. Automation/Scraping: Playwright (for dynamic content/cookies). AI: OpenAI API integration (GPT-4o). Server: Linux VP...
I need a developer to collect data from multiple public websites and deliver it in a clean, structured format. This is for legitimate data extraction from publicly available pages. I will share the target URLs and exact data fields with shortlisted candidates. Scope of work Scrape data from multiple public websites (details shared after shortlisting) Extract specific fields consistently and handle pagination/filtering where needed Normalize/clean the data (remove duplicates, consistent formatting) Export results to CSV/Excel/JSON (format to be confirmed) Provide a repeatable solution (script or small app) that I can run on demand Basic documentation: how to run it, how to adjust settings, where outputs go Quality requirements Reliable scraping...
Virtual Assistant E...Hubspot o Pipedrive). Capacidad para realizar scraping básico de datos (Google Maps, LinkedIn, sitios web). Excelente organización y atención al detalle. Dominio nativo o muy fluido del español. Inglés básico es deseable. Herramientas que debería conocer: Google Maps para scraping. LinkedIn Sales Navigator (altamente valorado). Aplicaciones de enriquecimiento de datos como Apollo, Snov, Hunter o Lusha. Google Sheets/ Excel a nivel intermedio. Herramientas de gestión de proyectos como Notion o ClickUp (deseable). Clientify o CRMs similares. Buscamos un profesional con mentalidad comercial, capaz de identificar oportunidades y contribuir activamente al crecimiento de nuestro pipeline de ventas, no...
I need help scraping data from a specific website. The focus is on collecting text content only; no images or contact details are required. I simply need all icon names from in a plain txt file, one icon name per row. sample: fa-heart fa-star fa-dinner etc timeframe: 12h max
...extension** that helps users: * discover and enrich business contact information (e.g. emails, company data) * save and organise leads * export data for outreach * upgrade to paid plans for higher limits and advanced features The extension is the **primary product**. A large standalone SaaS dashboard is *not* required for the MVP. --- ## Core Features (MVP Scope) ### 1. Chrome Extension (Manifest V3) * Works on supported pages (e.g. LinkedIn, company websites, public business pages). * Injects a clean UI panel or button. * User-triggered actions only (no background scraping). ---2. Lead / Contact Lookup * Extract basic public data from the current page (company name, domain, role, etc.). * Enrich data via third-party APIs (email discovery / ver...
...credentials required), submit a single text input as the search parameter, wait for the results, then harvest every HTML table that appears. Each table must be written to its own worksheet inside the same .xlsx file. Any images returned alongside the tables should be downloaded to a local folder the program creates during the run. Because the tool will run unattended on a server, everything—navigation, scraping, Excel generation, and image download—has to execute in the background without opening a visible browser window. JavaScript or REST calls that the site already exposes are fair game if they help keep the interaction truly headless. Deliverables • Visual Studio-ready C# source (targeting .NET 6 or later) with clear, in-line comments • A brief REA...
...Import Project Type: Data Scraping, Automation, Supabase Integration Delivery Timeline: ---- Days 1. Project Overview I am building an escort directory called Rosey using and Supabase (Postgres). I need a fully automated scraper and data pipeline that extracts all profiles from a target escort directory ( or ) and imports them directly into Supabase. This is a fresh data import , the scraped data becomes the canonical dataset for the platform. After initial setup, the system must run autonomously every hour to reflect new profiles, profile updates, and removed profiles from the source site. 2. Core Requirements - Scrape all profiles from the target directory ( or ) - Extract rich, structured profile data - Insert / upsert prof...
I have a set of websites whose data I need to capture automatically, and I want the whole process built as a reusable Apify actor. I will share the exact URLs, the fields to be collected, and the desired output format once we agree to proceed, but the common theme is structured extraction (think product specs, profile info, or similar). Here’s the outcome I’m expecting: • A clean Node.js actor that runs on the Apify platform, uses the latest Apify SDK, and follows best practices for request queuing, proxy rotation, and error handling. • Configurable input schema so I can plug in new target URLs or tweak search parameters without touching the code. • Output saved to an Apify dataset (JSON/CSV) and pushed to my Google Drive via webhook on each succe...
I need all relevant text pulled from a specific website and delivered in a clean, structured file. The source is entirely online—no PDFs or emails—so the job focuses on crawling or scraping the site, capturing every piece of visible textual content I specify, and returning it in a machine-readable format. I’m flexible on the final file type; CSV, Excel, or JSON all work as long as the fields are clearly labeled and easy for me to manipulate later. A small sample first will help confirm we’re on the same page before you run the full extraction. Please use whatever stack you prefer—Python with BeautifulSoup or Scrapy, JavaScript with Puppeteer, or a tool that suits the task best—just be sure to respect and provide the code so I can rerun the ...
...workflows Strong debugging and problem-solving skills under real-world constraints Background in one or more of: Trading systems Marketplaces AdTech / Fraud systems Load testing or simulation platforms Gaming or anti-cheat systems Familiarity with behavioral modeling or synthetic user generation Strong opinions on system design — and the ability to defend them What this is not Not a simple scraping task Not a junior automation role Not a basic CRUD application This is a senior engineering role requiring architectural thinking and production-grade implementation. Engagement details How to apply Please include: A short description of similar systems you’ve built (focus on scale & complexity) We need a full CV, without a CV your application will b...
I have a backlog of information that must be keyed in by hand—no scraping or automated pulls—so I need a detail-oriented freelancer who is comfortable with repetitive, precise work. You’ll receive source documents in PDF and image form, then enter the data into the Excel template I provide, following the field order exactly. If you also happen to edit educational videos, let me know; that skill could become useful in a later phase, though the immediate priority is 100 % manual data entry. Deliverables • Completed spreadsheet with every field populated and spell-checked • A quick note highlighting any ambiguities or missing values you encountered I value accuracy over speed, but a fast turnaround is still appreciated. Let me know you...
I need a clean pull of every location listed on For each branch please capture: country, state, complete address, service type, phone number, and email address. The final deliverable is a single Microsoft Excel workbook containing one sheet only. All columns should be clearly labelled and the range converted to an official Excel Table so I can apply native filters instantly. No additional filtering is required on your side; just be sure the table structure supports easy filtering by any column once I open the file. Accuracy matters more than speed—every location on the site has to be included and the contact details must match what is shown online. When you hand over the file I will spot-check a sample of entries against the live site to confirm completeness and correctness bef...
============================================================ ROLE: REAL ESTATE LISTINGS DATA & QA SPECIALIST (REMOTE) ============================================================ THIS IS NOT A GENERAL VA ROLE. This role is focused on structured, rule-based review and upload of real estate property listings. Accuracy and consistency matter more than speed. ------------------------------------------------------------ WHAT YOU WILL DO ------------------------------------------------------------ - Review and upload property listings using provided data - Verify price, location, property type, specs, and images - Identify missing, inconsistent, or suspicious information - Flag duplicate or low-quality listings - Follow written SOPs exactly (no improvising) - Record clear QA...
...one per client, each containing a minimum of 300 decision-maker leads (900+ total). • All companies must be UK-based—London and surrounding areas only. • Contact focus: operations or logistics decision-makers. No recruiters, no agencies, strictly no duplicates. Data points required for every record – Direct phone number (mandatory) – Work email address – Job title – Company name – City / location Process 1. Apply the exact Apollo filters I will share. 2. Export the resulting contacts—no extra scraping or manual research. 3. Double-check for duplicates across all three spreadsheets before delivery. Deliverables & turnaround • Three Excel or CSV files, clearly labelled by client, each...
...admin (daily) Clean, mobile-friendly table 4. Digital Product Sales Use Easy Digital Downloads (already installed) Sell business plans (PDF) Bundles + Full-Access option User accounts: Login Purchase history Re-download access 5. Payments Stripe (primary – global users) Paystack (secondary – Africa users) 6. General Requirements Clean, professional fintech UI Mobile-friendly No scraping No WooCommerce Simple admin workflow Basic handover instructions...
...For each lead, I must receive: • Contact details – at minimum the decision-maker’s name, direct email, and phone number • A clear indication of the business type or industry category • Website link (if outdated). You’re free to use whichever channels work best—email outreach, social media prospecting, telemarketing, data-scraping tools, LinkedIn Sales Navigator, Apollo, ZoomInfo—so long as the information you provide stands up to verification. Please supply the data in a clean spreadsheet (CSV or XLSX) ready for import into my CRM. I’ll sample the file on delivery; leads that bounce or prove inaccurate above a small tolerance will need replacing before final approval. When you respond, let me know: • The...
I want a Telegram bot that can reliably extra...Property: <Title> Unit No.: <unit_number> Client: <client_phone> Owner: <owner_phone> Source: <URL> Key points • No reliance on the Bayut or Propertyfinder APIs—pure scraping with your preferred stack (Python, Node, Playwright, Selenium, BeautifulSoup, etc.). • Handle anti-scraping tactics gracefully (rotating headers, proxies, captchas if they appear). • Keep response time reasonable so a conversation still feels instant. • Deliver clean, well-commented code plus a quick guide for deploying the bot on a VPS or Docker image. Acceptance will be a short live demo in Telegram showing the bot uncovering the three data points from at least two different ...
I need help compiling a clean, well-structured spreadsheet of ...same pages, including email addresses, phone numbers, and any listed location details. Requirements • Manual collection only; no automated scraping tools that violate terms of service. • Record each source URL alongside the data so I can verify accuracy. • Maintain consistent formatting in Excel or Google Sheets—one row per entity, separate columns for each data point. • Double-check spelling, remove duplicates, and flag anything that looks incomplete. Deliverable A finalized spreadsheet ready for immediate use and a brief note summarizing the total records gathered and any issues encountered. If you’ve handled similar online-source data entry proj...
...or XBRL when available) and label the files clearly. • Build a spreadsheet that lists each filing with the key metadata mentioned above. Acceptance The work is complete when I receive a zipped folder containing the documents and an error-free spreadsheet that cross-checks back to the live EDGAR links. Familiarity with tools like Python-sec-edgar-downloader, EdgarSearch API, or even manual scraping through the SEC site is welcome—use whatever approach you prefer as long as the output is accurate and reproducible. Turnaround is flexible but please outline how long you need per 100 filings so I can plan the next milestones....
...that I can later extend it to press releases or historical data if required. Here’s what I expect: • A script (preferably in Python 3 using requests / BeautifulSoup or Selenium if necessary) that accepts a plain text list of symbols, checks each page once per day and downloads any financial report that is not already saved. • Folder or filename logic that organises the PDFs by ticker and date so nothing is overwritten. • A simple log or CSV that records the timestamp, ticker and URL of each file fetched, plus any errors. • Clear instructions so I can schedule the job with cron / Task Scheduler and adjust the ticker list myself. When you respond, please focus on your experience building similar web-automation or scraping tools—link...
...on web scraping, data extraction, and data cleaning**. This is **NOT** a large system or customer-facing role. The work consists of: * Small, clearly scoped Python scripts * Web scraping (HTML, PDFs, APIs) * Data cleaning and transformation * ETL-style utilities All work is: * Async-first * Internal tools only * Clearly scoped with written requirements This is **ongoing contract work**. Strong performers may receive long-term work. --- ### What You’ll Be Doing * Build Python scripts to scrape public websites * Parse HTML, JSON, CSV, and PDF files * Clean and normalize messy real-world data * Write clear, maintainable utility scripts * Deliver working code (not just prototypes) --- ### Required Skills * Strong Python fundamental...
...large-scale listings, performance optimization, and legal-safe data handling. SCOPE OF WORK - Customize the ListingPro theme for education listings - Modify listing fields to support institution type (college, university, school, coaching) - Add ownership types (government, private, deemed, trust) - Configure courses and streams - Remove unused default ListingPro fields (restaurants, events, etc.) SEARCH AND FILTERS - Configure filters for city, state, course, ownership, and institution type - Ensure filters are optimized for large datasets (50,000+ listings) - Avoid heavy meta queries and performance bottlenecks LEGAL AND COMPLIANCE REQUIREMENTS (MANDATORY) - Add a global website disclaimer - Add a per-listing disclaimer - Do not scrape data, logos, or images from othe...
I need a reliable scraping solution that collects every open position from ten job-board and company-career sites in one specific country. I already have the full URL list and will share it right after kickoff. Scope • Write and schedule a separate scraper for each site, then funnel all results through a common pipeline. • Capture exactly three fields per post: Job title, Company name, and the full Job description. • Deduplicate across all sources so each role appears once, even if multiple platforms advertise it. • Store the clean data in either – a MariaDB database with a clear schema, or – well-structured JSON files (your choice; let me know which suits you best). • Provide a lightweight webview that lets me search, fil...
...experienced AI-powered web scraping specialist to build and operate a system that can collect thousands of U.S.-based b2b business leads per day from public directories such as: - Google My Business / Google Maps - Yelp - Better Business Bureau (BBB) - Similar public business listing platforms Each day, we will provide: - One U.S. city (changes daily) - A fixed list of industries (industries stay the same every day) Your job is to scrape ALL matching businesses in that city across those industries, at scale, and deliver clean, call-ready datasets for our sales team. Each record must include the following: - Business Name - Industry - Address - City - State - ZIP Code - Website URL (if available) - Phone Number -Business Owner (if possible) Accuracy matters — this data...
...styles, etc.) For each selected artwork: Retrieve key data (artist, title, medium, size, estimate, photos, auction house, sale date) Query ChatGPT (or OpenAI API) to: Assess artistic and market interest Compare estimates with historical auction results found online Evaluate whether the artwork appears undervalued or attractive Generate a daily report (email, PDF, or dashboard) summarizing: Selected artworks AI-based opinion (“interesting / neutral / not interesting”) Short justification for each recommendation Technical Expectations Preferred stack: Python, APIs, web automation/scraping (Playwright, Selenium, etc.) Integration with OpenAI / ChatGPT API Clean, well-documented, and maintainable code Respect of reasonable scraping limits and ...
We are looking for a freelancer to build a database of 1,000 active iGaming websites (casino / sportsbook / betting operators). Scope of Work: You will identify and collect 1,000 unique, live iGaming operator websites and enter them into our provided form/database. Each record will have several different data points, including but not limited to: - Website URL - Contact emails - Company / Brand Name - Country / Jurisdiction (where the operator is based or licensed) - Website Languages (select all that apply) - Availability of Providers Important Guidelines - Only live, operational websites (no affiliates, review sites, or news portals) - No duplicates (we control this in the record entry form) - We prefer speed and volume over excessive research - Do not spend significant time try...
I need to build a reliable, well-structured lead list and I already know exactly what it should contain. The task is to extract contact information—email addresses, phone numbers and full mailing addresses—from three sources: company and organisation websites, their public social-media profiles, and well-known online directories. I expect the data to be gathered with a solid scraping workflow (Python, Scrapy, BeautifulSoup, Selenium or an equivalent stack is fine) and then verified so that bounced emails and dead numbers are kept to an absolute minimum. Deliverables • One CSV or Excel file with separate columns for name, company, job title, email, phone, street address, city, state, ZIP/postcode, country, source URL and date collected. • No dup...
I’m building an internal dashboard that tracks every maritime-related vacancy published online and I need a robust, repeatable scraping solution. My scope is broad: I want coverage of all major job boards that regularly post roles for the shipping, offshore, and port-services sectors. Because I don’t have a definitive site list yet, I’ll count on you to recommend the platforms you know are rich in maritime listings—think the big global boards as well as any niche maritime career portals you’re aware of. From each posting, I must capture the following fields: • Job title and full description • Company name plus location (city, state/region, country) • Employment type and any salary or rate information available Your scraper shoul...
I need every bicycle, accessory, and clothing item currently sold on imported into my own e-commerce site. Each listing on my end m...correct titles, descriptions, prices, and image galleries. • Products appear in the appropriate categories that mirror the structure on Bike24 so customers can browse naturally. • I receive clear instructions or an automated routine that lets me refresh the data in future without writing new code. If you have handled large-scale catalogue migrations or have experience with web scraping tools, WooCommerce importers, Shopify APIs, or similar platforms, that background will help us move quickly. Let me know how you plan to tackle the data extraction, how long the first full import will take, and any tooling or access you&rsq...
Description: - We are looking for an experienced Data Scraping / Web Scraping expert. - We will share the industry name, and the freelancer should: - Suggest suitable websites/sources to scrape - Suggest countries/regions that can be covered - Share estimated data volume & approach - After approval, the freelancer will scrape and deliver clean, structured data. Data Required (example): - Company name - Location - Contact details (email/phone/website – if available) Requirements: - Proven experience in data scraping - Knowledge of Python, Scrapy, Selenium, APIs, etc. - Ability to scrape multi-country data (based on feasibility) Deliverables: - Data in Excel / CSV / Google Sheets - Basic info of sources used ...
...price and automatically compare it to the historical data of recent actual sales found on secondhand marketplaces (eBay, Wallapop, wallapop). The VPN rotation switches between different Surfshark server configuration files (.ovpn files) using the same account credentials, which changes the IP address by connecting to different VPN servers. Project Status: 95% Complete – Final Production Readiness Tasks ✅ Production Server: Deployed on DigitalOcean ✅ Real Data Scraping: Live Vinted & Wallapop marketplace data ✅ Professional Dashboard: Bootstrap 5 web interface ✅ Database System: MySQL with complete schema ✅ Email Alerts: SMTP notification system ✅ Safe Scraping: SSH-preserving method (15-30% success rate) ✅ Export Functions: CSV/Excel da...
Senior Full-Stack Engineer / Technical Lead – Data Ingestion, Web Scraping & AI Systems (Long-Term) Job Description We are a fast-scaling real estate investment and technology company building a data-driven platform focused on sourcing, analyzing, and tracking commercial real estate opportunities across the U.S. We are looking for a senior-level engineer / technical lead who can help us own and scale our existing platform, improve reliability, and build robust data ingestion systems. This is not a short-term task — we are looking for someone who wants to grow with the project and eventually help lead a broader technical team. This role is ideal for someone who is hands-on, systems-oriented, and proactive, not someone who needs detailed instru...
...businesses shown there. Here is the scope in plain terms: • Extract every available address field (street, city, state/province, postal code, country if present). • Organise the data in a clean, comma-separated CSV with consistent column headers. • Validate that the final count reaches 20,000 unique records; no duplicates. Timing & payment I’m paying a flat $20 and I need the finished CSV back within 12 hours of project acceptance, so please only respond if you already have the tools and experience to hit that deadline. Acceptance criteria • 20,000 unique mailing address rows delivered on time. • Data matches what appears on the live business directory site. • File opens without errors in Excel or Google Sheets...
...modular, well-commented codebase Strong technical documentation User & Authorization Structure Admin Panel Admin users with role-based access control Ability to create and manage agency (company) accounts System-wide configuration and monitoring Agency Users Ability to manage multiple applicants Support for different visa types and profiles Ability to trigger automated booking jobs using saved data Automation & Intelligent Behavior Periodic and automated checking of VFS appointment availability Ability to stay logged in during high-demand periods Immediate booking when a slot becomes available Human-like interaction techniques: Mouse movement simulation Typing delays and patterns Randomized wait times Optional virtual keyboard usage Secure browser context isolation Rob...
I need all publicly available customer-facing email addresses extracted from a list of e-commerce websites that I will supply once the project begins. Please crawl only the domains I provide, respect where possible, and avoid triggering any rate limits or security blocks—rotating proxies or headless browsing with tools such as Python, Scrapy, BeautifulSoup, Selenium, or similar is fine as long as the result is reliable. Deliverable • One clean, de-duplicated CSV file containing the harvested email addresses, ready for direct import into my CRM. Acceptance criteria • Every email must originate from the target e-commerce domains. • No duplicates, placeholders, or obviously invalid addresses. • File encodes as UTF-8 and opens without warnings in Excel/Goog...
...handled programmatically—token refresh, headers, or cookies—so the run stays unattended. • Pagination or “next-page” parameters followed until no records remain, with graceful back-off to respect any rate limits. • For every case, text fields saved to a consolidated JSON or CSV file, and each PDF stored locally (or optionally pushed to S3) with a predictable filename that is referenced in the data file. • Robust logging, retry logic, and a simple resume flag so I can rerun without duplicating previously collected items. I will provide the endpoint list, example credentials, and a small expected output sample once we start. The work is done when I can clone the repo, run `npm install && node `, and watch it finish without error...
Hello, I am looking to work with an AI Research & Automation Specialist who can help me build an AI-driven system for structured data discovery, intelligence mapping, and decision-maker identification. This is not a traditional sales or manual research role. The focus is on using AI tools, prompt engineering, and automation to extract, structure, and continuously improve high-value data with minimal human dependency. Due to the nature of the data involved, the selected individual must be willing to work under a formal contract with strict data confidentiality and non-disclosure obligations. Core Objective To design and operate an AI-powered intelligence workflow that can autonomously discover, organize, and prioritize: * Government and public-sector engine...
I need a small Python script that can visit a specific website, pull the information I’ll define (for example prices, status messages, or any visible text blocks), run a simple evaluation on what it finds, and immediately push the result to my Telegram account or channel. I’m not fixed on the scraping stack as long as it’s reliable and easy to maintain. If you have suggestions for handling dynamic content or rate-limiting, let me know in your proposal.
...filter I care about right now) • cycle through each File Number that appears • open the property page that pops up in a new tab or frame • copy the table cells for Owner Name, Father/Husband Name, Correspondence Address and Share. Doing this by hand is no longer practical, so I want you to recreate that exact sequence in code or with a scraping tool that can handle the site’s dynamic dropdowns and postbacks. This is intended as a one-time data pull, but I’m not opposed to receiving the script as well if it makes future updates easier—I’m simply not sure yet whether I’ll need ongoing automation. Deliverable acceptance will be straightforward: when I open the spreadsheet I should see every CP and CPL record, one row per fi...
I need someone who already holds...to informatique (inks, PCs, etc.), construction tools, and raw materials. • Export every available column Volza provides—volumes, values, supplier names, buyer names, dates, ports, and any other data fields you see. • Deliver the information as a clean Excel file; no further segmentation is necessary. I simply want the complete, unaltered rows for the selected HS codes. Workflow & cadence 1. First delivery: complete dataset for 2024 and the year-to-date portion of 2025. 2. Ongoing: refresh the file every six months so the database stays current. A solid grasp of data scraping or bulk export within Volza is essential. If you have the subscription and can automate the pull, we can turn this into...
...AI step that writes a concise English summary and produces an accurate Arabic counterpart. Everything is then pushed to a simple public page or email like google alerts showing URL, title, and the two-language summary. Deliverables • n8n workflow file (.json) with annotated nodes (I will give you access to my n8n account if needed) • Any custom JavaScript/TypeScript or HTTP requests used for scraping or API calls • Auto-generated web page (or CMS integration) ready to publish, styled minimally • README with setup steps, credentials needed, and how to add or swap sources Acceptance criteria • Workflow runs autonomously every 12 hours without manual restart • Pulls from AI-recommended feeds plus my provided list • Produces at least ten ...
...developer with strong skills in web scraping, background task processing. - Project Scope The goal is to build a background scraping system and display the collected data in a Django application. - Key tasks 1. Run web scraping tasks in the background on an Ubuntu VPS, using proxies and basic anti-bot techniques. 2. Store the scraped data in a PostgreSQL database connected to the Django project. 3. Create a simple Django page to display the scraped data. - Requirements * Proven experience with Django and react * Experience with web scraping and anti-bot measures * Background task processing (Celery, cron, or similar) * PostgreSQL integration * Ability to deliver a working, completed solution - Acceptance Criteria * Scraping...
I need a developer to collect data from multiple public websites and deliver it in a clean, structured format. This is for legitimate data extraction from publicly available pages. I will share the target URLs and exact data fields with shortlisted candidates. Scope of work Scrape data from multiple public websites (details shared after shortlisting) Extract specific fields consistently and handle pagination/filtering where needed Normalize/clean the data (remove duplicates, consistent formatting) Export results to CSV/Excel/JSON (format to be confirmed) Provide a repeatable solution (script or small app) that I can run on demand Basic documentation: how to run it, how to adjust settings, where outputs go Quality requirements Reliable scraping...