Selenium webdriverJobs
WebDriver VB6 NotariaReabiertas + Updates (NotasSimples+RedSara)
This is an ongoing project and most likely it will turn into a long term business relationship if things play out accordingly. MUST have: Python Selenium/Scrapy Django/Flask HTML/CSS/Boostrap JS MySQL, Postgres, Sqlite Redis, mongodb AWS, Heroku GIT + zenhub TDD, SOLID Nice to have: -React/Redux -CI, CD -Graphql -Elastic search Please show your past experience in your proposal PD: state what passions you about working as a freelancer in less than 10 words. all details will be provided to the few selected throu DM. Best Regards!!
...successful candidate How it works: • You refer a candidate who needs interview support • Candidate confirms and takes the service • You receive $30 per successful referral • No limit on referrals – unlimited earning potential Technologies Covered: • Software Development: React, Node.js, Java, Python, .NET, Full Stack • Cloud & DevOps: AWS, Azure, GCP, Docker, Kubernetes • Testing: Automation, Selenium, Cypress, Manual Testing • Data Roles: SQL, Data Engineering, Data Analyst, AI • Engineering: Controls Engineering , Automation, Embedded, Electrical • Enterprise Tools: SAP, Salesforce, ServiceNow, Oracle Why partner with me: • Transparent process • Fast payout after successful referral • Long-term partne...
...post) unless properly bound to browser session Maintain real browser fingerprint consistency Preferred Technical Approach: Selenium / SeleniumBase integration Chrome DevTools Protocol (CDP) usage Network interception or replay In-browser JavaScript execution (fetch / XHR override) Proper handling of dynamic request signing if applicable Goal: To programmatically trigger the same successful API call that the browser makes, while preserving complete browser session authenticity. Deliverables: Fully working implementation Clean and optimized code Explanation of approach used Must be tested and verified against protected endpoint Required Skills: Python Selenium / SeleniumBase Chrome DevTools Protocol (CDP) Browser Network Debugging JavaScript (fetch / XHR ...
...unless properly bound to browser session Maintain real browser fingerprint consistency Preferred Technical Approach: Selenium / SeleniumBase integration Chrome DevTools Protocol (CDP) usage Network interception or replay In-browser JavaScript execution (fetch / XHR override) Proper handling of dynamic request signing if applicable Goal: To programmatically trigger the same successful API call that the browser makes, while preserving complete browser session authenticity. Deliverables: Fully working implementation Clean and optimized code Explanation of approach used Must be tested and verified against protected endpoint Required Skills: Python Selenium / SeleniumBase Chrome DevTools Protocol (CDP) Browser Network Debugging JavaScript (fetch / XHR ha...
...content in a clean CSV or Excel file with clear column headings; if you prefer a database export, let me know and we can adjust. • Include the finished script or notebook so I can rerun the extraction later. Accuracy and formatting matter more to me than sheer speed, so please allow time for basic validation before handing over the files. If you normally work with Python (BeautifulSoup, Scrapy, Selenium) or similar tooling, that’s perfect, but I’m open to alternative stacks as long as the output meets the same standard. When you reply, briefly outline: 1. The scraping approach and libraries you’d use 2. Any anti-blocking measures you apply for public sites 3. A realistic timeframe to capture, clean, and hand back the data I’m ready to start a...
Project Title: WhatsApp Transfer Verification & Auto Receipt System Project Descriptio...representative on WhatsApp. - Mark the transfer as approved. 5. Automatically open my company accounting system (web-based). 6. Create and submit a Receipt Voucher (Cash Receipt Entry) with the transfer details. Technical Requirements: - Python or Node.js - WhatsApp API integration (or automation if API not available) - OCR for reading transfer receipts (Arabic + English) - Web automation (Selenium / Playwright) - Secure handling of credentials - High accuracy verification logic Deliverables: - Full source code - Executable version (EXE if Windows) - Setup documentation - Training on how to use the system Goal: Automate transfer confirmation and receipt entry to save time and reduc...
I need a reliable browser-based bot that logs in to my Amazon A to Z account, continuously monitors upcoming shifts and instantly books the ones that match criteria I will later define (location, start time, hours). The tool must operate inside a standard web browser—headless Chrome or a Selenium/Puppeteer script is fine—as long as it books faster than manual clicking and survives Amazon’s usual page refreshes, captchas, and timers. I will grant the bot full access to my account solely for booking purposes, so handling login securely (2FA, encrypted credentials, cookie reuse) is essential. A simple configuration file or UI where I can tweak preferred warehouses, shift lengths, and daily booking windows would be helpful, but speed and reliability come first. Deliv...
...publicly available sites. The focus is simple yet crucial: for every company you find, capture the homepage URL and a working email address. (ask for details in the sheet ) A completely ethical approach is non-negotiable—no gated content, no third-party lists, and no automated harvesting that violates site terms. I’m happy for you to use tools you’re comfortable with (Python, Scrapy, BeautifulSoup, Selenium, Google Apps Script, etc.) as long as you respect and rate limits. Email addresses must appear in plain text within the sheet; please avoid hyperlinks or HTML encoding. Deliverables • A Google Sheet populated with data • A short note on your collection method (manual, scripted, hybrid) so I can replicate or update the data in the future. ...
...location de bateaux. Pour y parvenir, j’ai besoin d’un workflow automatisé reposant exclusivement sur Google Maps – c’est la source retenue – capable de collecter, dédupliquer puis nettoyer les données avant de les mettre en forme dans un CSV directement importable dans le CMS Wix. Les coordonnées devront être fournies en degrés décimaux. Livrables attendus • Un script réutilisable (Python + Selenium, Scrapy ou équivalent) qui interroge Google Maps, gère le rate-limit et documente chaque étape de traitement. • Le fichier CSV final contenant, pour chaque base nautique, les champs suivants : Nom, Adresse complète, Ville, Département, Régio...
...búsquedas y resultados Entregables Código fuente completo Documentación de instalación y uso Documentación de endpoints de la API Ejemplo de request/response (texto e imagen) Deploy listo en servidor/VPS o guía de deploy Importante El sistema debe ser escalable y mantenible, ya que se planea agregar más tiendas con el tiempo. Perfil buscado Experiencia comprobable en scraping (Playwright/Puppeteer/Selenium) Experiencia creando APIs REST Experiencia en visión computacional / reconocimiento de imágenes (deseable) Ideal si tiene ejemplos previos de scraping e-commerce o monitoreo de precios Para aplicar Por favor enviar: tecnologías que usarías enfoque para búsqueda por imagen (tipo Lens) ejemplos...
I am looking to build a custom AI agent that manages my end-to-end job application workflow. The goal is to create a semi-au...new matched jobs • Review drafted applications • Approve / reject submissions • Track applied jobs • View outreach sent • See recruiter responses • Track interview updates Clean UI preferred. Minimal but functional. ⸻ Technical Expectations Open to suggestions, but likely stack: • Python / Node backend • AI via OpenAI API or similar • Automation using browser automation (Playwright/Selenium) • Email integration (Gmail API) • LinkedIn automation (must comply with platform limits) • Cloud deployment (AWS/GCP/Azure) • Simple web/mobile-friendly dashboard (React / Flutter / etc.) ...
...Python crawler that pulls public content from Twitter, Instagram and LinkedIn, covering text, image and video posts for any handle I feed it. Here’s the flow I have in mind. The script collects the raw post data (caption, hashtags, basic engagement numbers and, where accessible, image/video URLs) through whichever mix of libraries makes sense—Tweepy or Twitter API v2 for Twitter, Instaloader or Selenium for Instagram, and the official or unofficial LinkedIn API for LinkedIn. After normalising everything into a common JSON schema, the crawler should pass that dataset to an LLM endpoint (OpenAI or similar) and receive back a concise, structured report that includes: • Brand sentiment (positive / neutral / negative trends) • Key thematic buckets the bran...
Project Title Custom Lead Generation & Email Scraper Tool (Google, Yellow Pages, & B2B Directories) Project Description I am looking for an experienced developer to build a robust, high...Email Address (Must have a validation check to avoid "dead" emails) Phone Number (Optional but preferred) LinkedIn Profile URL (Optional but preferred) Export Functionality: Capability to export data into CSV or Excel format. Anti-Blocking Measures: Use of rotating proxies or delays to ensure the scraper isn't blocked by Google or directories. Technical Requirements: Preference for Python (Selenium, Scrapy, or BeautifulSoup) or a dedicated desktop application. User-friendly interface (even a simple CLI is fine, but a GUI is a plus). Fast processing speed with the abil...
Project Title Custom Lead Generation & Email Scraper Tool (Google, Yellow Pages, & B2B Directories) Project Description I am looking for an experienced developer to build a robust, high...Email Address (Must have a validation check to avoid "dead" emails) Phone Number (Optional but preferred) LinkedIn Profile URL (Optional but preferred) Export Functionality: Capability to export data into CSV or Excel format. Anti-Blocking Measures: Use of rotating proxies or delays to ensure the scraper isn't blocked by Google or directories. Technical Requirements: Preference for Python (Selenium, Scrapy, or BeautifulSoup) or a dedicated desktop application. User-friendly interface (even a simple CLI is fine, but a GUI is a plus). Fast processing speed with the abil...
My team has a fully-built web application and I need a fresh set of expert eyes to validate every line of code. The goal is to run a complete test cycle—functional, performance, and security—so we can ship with confidence. You are free to recommend and use whichever framework or tool best fits the job; I am open to industry standards such as Selenium, Playwright, JMeter, OWASP ZAP, or anything you feel will surface issues quickly and reliably. What matters most is deep coverage of the entire source code, clear defect reporting, and actionable suggestions for hardening and optimisation. I will provide the full repository, environment details, and access credentials once we agree on the engagement. In return, I expect: • A concise test plan outlining scope, assump...
I already have a curated list of LinkedIn profile URLs and need the key networking details moved into a single Google Sheet. For every profile, capture each person’s stated interests and list the five types ...(as it appears) • Interests (comma-separated) • Type 1 through Type 5 (verbatim wording) • Category tag (Industry experts / Potential clients / Collaborators) Accuracy of the text you pull is more important than speed, but I do expect the work to comply with LinkedIn’s terms and avoid triggering any scraping limits. If you prefer to work manually, that’s fine; if you script with Python, Selenium, or similar, just make sure the final output lands cleanly in my shared Google Sheet. I’ll review a small sample first, then green-light...
...it detects any difference in either the visible text or the images. The page updates unpredictably, so true real-time tracking is essential; a one-minute polling interval is already too slow for my use-case. A lightweight approach that respects the site’s bandwidth and avoids triggering blocks or captchas will be valued. I am open to whatever stack you favour—Python with BeautifulSoup or Selenium, Node.js with Puppeteer, or a compiled solution—so long as it is stable on a Windows environment and easy for me to tweak the target URL later. Notification method is flexible: an email is fine, but if you have a smarter suggestion (desktop toast, webhook, etc.) that can be set up quickly, I’m all ears. Please deliver: • The complete, well-commente...
I’m spinning up additional quality coverage for an active web and mobile project and need a seasoned tester who can switch easily between writing code-driven checks and running thoughtful manual sessions. Your day-to-day will include drafting concise test plans, building and maintaining Selenium-, Cypress- or Playwright-based suites, and verifying APIs with Postman or a similar tool. When something breaks, I’ll count on you to log reproducible defects and follow them through to closure, collaborating directly with developers in our Agile board so each sprint ships clean. Because delivery timelines are tight, I’m specifically after someone who already has strong automation foundations, understands functional, regression and end-to-end coverage, and can work indepen...
Title: Contract QA Automation Engineer – Onsite (3–6 Months) Project Overview: We are hiring QA Automation professionals for a short-term contract (3–6 months) to support active development projects. Location (Onsite/WFO): Udaipur | Jaipur | Bangalore | Bhopal Skills Required: Proven experience in automation testing Selenium / Cypress / Playwright or similar tools API testing knowledge Strong problem-solving skills Contract Details: Duration: 3–6 months Full-time engagement Onsite only (no remote option) Who Should Apply: Freelancers or contract professionals available for onsite work Immediate to short notice joiners preferred
...using Playwright, one of the fastest-growing and most powerful automation frameworks in the market. This comprehensive course is designed to take you from complete beginner to advanced automation engineer through hands-on projects, real-world scenarios, and framework development. Playwright Full Course - Complete Table of Contents Beginner Level • Introduction to Playwright • Why Playwright vs Selenium and Cypress • Supported Browsers and Languages • Environment Setup (Node.js, IDE, Installation) • Project Structure and Test Runner • Understanding Locators and Selectors • Basic Actions: Click, Fill, Type, Select • Assertions using Expect API • Handling Waits and Timeouts • Debugging with Inspector and Codegen • Running an...
...pages in real time or on a frequent schedule. • Apply NLP or other classification techniques to decide whether a posting is truly AI-related, then tag it by sub-domain (e.g. vision, NLP, MLOps, prompt-engineering). • Deliver concise, deduplicated listings to me through an in-app notification feed—no email or SMS required. For the deployment side I’m open to Python (Scrapy, BeautifulSoup, Selenium), Node, or any stack you are comfortable with so long as it is containerised and can run unattended on a small cloud instance. A lightweight web interface or Electron desktop app for the notification feed is ideal; you can suggest an alternative if it achieves the same user experience. Acceptance criteria 1. Agent successfully scrapes at least three live sou...
...captured each record logged beside the data. • Consistency matters: please apply uniform naming conventions (e.g., “FY2023 Gross Profit” instead of varying labels) and check subtotals or totals to be sure everything reconciles. • I’m flexible on the final file type—CSV, Excel, or Google Sheets all work—so let me know which you prefer or suggest. • If you automate with Python, BeautifulSoup, Selenium, or a comparable tool, great; just include the script so the process can be rerun later. A quick README explaining any inputs or environment setup is enough for me to replicate it. • Accuracy is non-negotiable. I will spot-check figures against the original web pages, so double-check before submitting. Once you deliver the dat...
I need end-to-end testing for my web application. The project code is already written, so you will be focusing solely on testing. Requirements: - Thoroughly test on Chrome - Ensure all workflows function as intended - Identify and document any bugs or issues Ideal Skills: - Experience with end-to-end testing tools (e.g., Selenium, Cypress) - Strong attention to detail - Familiarity with web applications and browser testing Please provide a brief overview of your testing experience and any relevant tools you plan to use.
I need a reliable solution that can pull data from LinkedIn and insert it straight into a database I specify. The core requirement is the automated transfer—once the tool fini...strings and add new profile URLs without touching the core logic Acceptance criteria • I point the tool at 10 test profiles; all mapped fields appear in the database exactly once with no missing characters • A command-line log or dashboard shows progress and any errors in real time • You hand over the source code, a README with setup steps, and a sample SQL schema Experience with Python (Selenium, Playwright or similar), LinkedIn APIs, and database integration will be highly valued. Let me know your preferred stack, any relevant past projects, and the estimated timeframe to del...
I need a straightforward Python script that signs in to a Hotmail / account with Selenium (or another reliable browser-automation library)
I need end-to-end testing for my web application. The project code is already written, so you will be focusing solely on testing. Requirements: - Thoroughly test on Chrome - Ensure all workflows function as intended - Identify and document any bugs or issues Ideal Skills: - Experience with end-to-end testing tools (e.g., Selenium, Cypress) - Strong attention to detail - Familiarity with web applications and browser testing Please provide a brief overview of your testing experience and any relevant tools you plan to use.
I have a public-facing website that I need scraped end-to-end. The site is open (no login), but the content is split across multiple pages, so your script will have to detect and follow pagination automatically. Here is exactly what I expect: • A clean, well-commented Python script (requests/BeautifulSoup, Scrapy, or Selenium—your choice) that visits every page, captures the required fields, and writes them to a neatly structured CSV. • The final CSV containing all rows pulled from the site. • A short README that tells me how to run the script and change the target URL or output path if needed. Code quality matters to me: no hard-coded absolute paths, clear variable names, and graceful error handling so the run doesn’t stop if a single page fa...
I...framework into a CI pipeline. • Walk-throughs of flaky-test triage, mocking external dependencies, and debugging failures that only show up in complex environments. • Short take-home exercises or sample repositories that reinforce each lesson, plus code reviews so I know I’m applying the patterns correctly. Although my main interest is integration testing, I’m flexible on specific tooling: Selenium, NUnit, SpecFlow or whichever stack you feel showcases best practices in C# automation. The important part is understanding why a tool is chosen and how to extend or swap it later. Please outline your preferred tools, how you normally structure a learning path, and roughly how many hours you expect we’ll need to reach a self-sufficient framework and...
...support release cycles. • Maintain test documentation and contribute to continuous improvement of QA processes. Required Skills & Qualifications Core QA Skills • 3+ years of experience in software testing, preferably in cybersecurity or networking domains. • Strong understanding of QA methodologies: black-box, white-box, and grey-box testing. • Experience with test automation frameworks (e.g., Selenium, PyTest, Postman). • Familiarity with reverse proxy tools (NGINX, HAProxy, Envoy) and network traffic analysis. • Hands-on experience with Linux environments and shell scripting. • Graduate from a Top 50 engineering college in India as per NIRF 2025 ranking. Cloud & Security Awareness • Testing in cloud platforms (AWS, Azure, GCP) ...
...implement a robust test automation architecture with Maven/Gradle and CI/CD pipelines Page Object Model (POM) or similar design pattern reporting (Allure/Extent Reports or similar) reusable and maintainable test scripts documentation for setup and usage Required Skills: experience in Playwright with Java understanding of Selenium/automation concepts with TestNG/JUnit in CI/CD integration (Jenkins/GitHub Actions, etc.) of Git version control in framework design and best practices Nice to Have: automation experience knowledge testing exposure Duration: Short-term (with possible extension) Share your previous Playwright (Java) project
...CSV should include original variables like organization name, state and zip even though that data was not used in the scraper. The script must perform the following steps for each URL in the input list: 1. Input: Read a list of URLs from a provided CSV file (single column of URLs). 2. Navigation/Rendering: Visit the URL (handling redirects is essential). The use of a headless browser (like Selenium/Puppeteer) or an advanced HTTP library is preferred, as some websites may load the footer content dynamically via JavaScript. 3. Targeted Scanning: Scan the HTML source code of all pages found in the sitemap, specifically looking for the presence of a specific link. 4. Output Logic: - If the link is found, record the identified vendor. - If no vendor is explicitly identified, ...
I need the entire history of a specific Facebook Group captured—every post along with all associated comments. I’m ...with working links to the images and videos placed in clearly named folders. I don't want folders or links. Just one huge continuous page that has everything. This is for a court case and I have to give this to the other side. I want them to have to scroll through however many hundred pages there are. Just as if they were actually on FB. Please outline: • your scraping approach (Python + Selenium, Go, node-puppeteer, etc.), • how you’ll handle media downloads and folder structure, • estimated turnaround time. I’ll review a short sample export before we proceed with the full run to confirm the layout meets my ...
...need a Selenium-based solution that runs reliably on Windows and opens Google Chrome to simulate human visits to LinkedIn (and occasionally other) profile URLs listed in a Google Sheet. For each URL the program should: • Pull the next unused link from the sheet • Load the page in Chrome, wait a random time between 20 seconds and 3 minutes • Apply truly randomized scrolling patterns while the profile is open so behaviour looks organic • Fire a webhook the moment the visit completes, passing back any ID or payload I define so our CRM reflects the touch instantly Configuration items such as Google Sheet ID, webhook endpoint, minimum/maximum dwell time, and daily visit caps should live in a simple file I can edit without touching code. A short README on ...
...the comment text, number of comments, likes, reposts/shares, the post date and any other readily available metadata (author handle, follower count, post URL, media links, etc.). Accuracy is critical because the data will feed a trend-analysis dashboard later. Please build the workflow in a way that respects rate limits and login requirements: if you intend to use official APIs, private APIs, Selenium, Scrapy, Playwright, or headless browsers, spell that out so I know how sustainable the solution will be. The final hand-off should include: • A clean, well-commented reusable script (Python preferred) • A short README explaining environment setup, keyword input format and how to extend to new regions • The full export in CSV so I can validate before sign-off I...
I need a r...I can decide the exact days and times each role goes out • one-click “auto post” that instantly publishes a job when I hit the button The posting frequency isn’t fixed; some days I may blast several openings, other weeks none at all, so the scheduler has to respect whatever plan I set. Use whichever method makes the process most stable—headless browser automation (Puppeteer, Playwright, Selenium) or direct API calls if Placementindia offers them. The final package must run on a standard VPS, expose clear logs/errors, and allow easy editing of job templates. Please share links or short videos of similar bots you’ve built so I can gauge robustness. Once I can log in, queue a job, and watch it appear live without intervention, I&rs...
...Primitiva: €1.00 El Gordo: €1.50 Dynamic AI suggestions: recommend bet quantity based on backtesting, current jackpot and balance, using Monte Carlo for EV (prioritize positive EV scenarios). 2.5. Login and Bet Execution Automation Never store login/password (not even encrypted). Prompt user for credentials (email/NIF/NIE + password) every login, using secure masked fields. Login process: Selenium (headless Chrome/Firefox), access , fill form, submit, verify success (post-login elements), handle sessions/cookies. Bet execution: navigate to lottery page, generate AI numbers, add exact user-requested bets to cart (batches for high volumes), support rules of each lottery. Finalization: show full summary (lottery, bets, cost, numbers); require
I need a small, Windows-friendly Python script that will open a real browser with Selenium and wipe large batches of content from my X (Twitter), Facebook, and Instagram accounts. Because my X account sits on the free API tier I keep running into 403 errors, so this project must rely solely on browser automation—no official APIs or paid third-party tools. Here’s what I’m after: the script launches from the command prompt, asks for (or reads from a .env) my login credentials, signs in, and then iterates through all visible posts, tweets, and reels, deleting each one until none remain or until it hits an optional stop condition such as a date or a post count I can set. A simple console printout like “Deleted tweet #42” is enough for logging; I don’...
IM TYRING TO RUN THE ATTACHED JPNY SCRIPT TO GET INFO FROM A WEBSITE BUT I CANT UNDERSTAND IT DOESN'T WORK. I NEED THIS SCRIPT TO BE FIX + PAGINATION TO FETCH AROUND 2400 RECORDS FOR YELLOWPAGES I ONLY USE JUPYTER
...market research. The job centers on extracting selected data points from public web pages, transforming them into a clean, structured format, and making them available for analysis every 24 hours. Here’s what I need you to handle from end to end: • Source acquisition – fetch HTML from the URLs I provide, even when content is hidden behind JavaScript (a headless browser such as Playwright or Selenium is fine). • Parsing & cleansing – pull the specific fields I’ll list (product name, price, SKU, availability, and a time-stamp), remove duplicates, and standardize values. • Storage & delivery – load the daily output into my PostgreSQL instance; if you prefer Parquet or plain CSV that’s acceptable as long as it’s a...
...information should be organised into a clean CSV file—one row per page—with columns for page URL, full body text, image file names, and link destinations. Please download the images themselves as well and bundle them in a separate folder (a simple ZIP is fine); the CSV should reference the exact filenames so everything lines up. I’m happy for you to use Python with BeautifulSoup, Scrapy, Selenium or whichever stack you prefer, as long as the final output meets these acceptance criteria: • Complete CSV containing text, image names, and link URLs for each page • All images successfully downloaded and accessible via the filenames listed in the CSV • No duplicates or missing pages from the target site * Images need to be sorted for each l...
I am looking for a Python developer to create a simple and focused scraper script for Facebook Mar...• A file containing all product URLs for that seller • File format: TXT or CSV • Handle infinite scrolling to load all products Technical Requirements: • Python • Selenium or Playwright • Experience with dynamic websites • Clean, runnable, and well-structured code Important Notes: • No filters required (no country, city, or keywords) • No data is needed other than product links only • Manual login can be used if required Budget: Open — to be discussed based on experience and quality When Applying, Please Include: • Any previous experience with Facebook Marketplace scraping • The tool you plan to use (...
I have a data-analysis pipeline that relies on a steady flow o... • Payload: high-resolution image files plus a CSV/JSON map linking each file to product ID, title, price, and category text that you extract during the same run. • Scale: thousands of products per crawl; a resumable approach is essential so partial failures don’t force a full restart. • Frequency: I’ll trigger the crawl weekly, so reusable code is a must. I’m happy with Python—Scrapy, Selenium, Playwright, or a headless solution of your choice—as long as it respects the site’s anti-bot measures and keeps requests polite. Please include a brief outline of how you’ll handle pagination, lazy-loaded images, and rate limiting. Let me know your proposed stac...
...mandatory declarations and any digital signature steps must all be handled by the system before it attempts the final submission. A short, clear dashboard that shows “ready”, “errors found” or “submitted” status for each tender would be ideal so I can intervene only when something is missing. Deliverables I must see to accept the project: • Source code and install guide (preferably Python with Selenium / Playwright or similar RPA layer, but I’m open if you can justify another stack) • A configuration file where I can add new government portals without touching core code • Automated validation rules that stop a submission if any mandatory field or attachment is missing • A post-submission PDF/CSV log summarisin...
IM TYRING TO RUN THE ATTACHED JPNY SCRIPT TO GET INFO FROM A WEBSITE BUT I CANT UNDERSTAND IT DOESN'T WORK.
I need a small automation script that periodically checks item availability on the Bigbasket website and pings me on Telegram the moment any of the tracked products come back in stock. You are free to choose the underlying tech stack (Python + Requests/BeautifulSoup, Selenium, Playwright, or a headless browser of your choice) as long as it works reliably with Bigbasket’s current site layout and protects my account from rate-limit blocks or captchas. The flow I have in mind is straightforward: I feed the bot a list of product URLs (or SKUs). It runs on a schedule I can change—every few minutes during peak shortages, maybe every hour otherwise—grabs the stock status, and fires a concise Telegram message whenever the status flips from “Out of Stock” to &l...
...workbook. Please crawl the entire site, not just a few sections, and return each number alongside the key profile details that make the data usable at a glance—name, profile URL, and any other easily captured identifiers shown next to the number. A clean .xlsx with one row per profile, no duplicates, and clearly labelled columns is the only deliverable I’m expecting. If you prefer Python, Scrapy, Selenium, Beautiful Soup or a comparable stack, go ahead; I’m interested in results, not the specific toolset, as long as the script can be rerun later should the site content change. Before delivery, double-check that: • every row contains a valid phone number and url • no pages on the site were skipped • the sheet opens flawlessly in the latest des...
Quiero contar con un archivo .xlsx que contenga las 12 728 filas completas de la tabla pública que aparece en la web de INDECOPI (Perú). El sitio sólo muestra 10 regis...formato estándar, sin filtros, tablas dinámicas ni otras funcionalidades añadidas. Yo te facilitaré la URL exacta y los pasos de navegación para que ubiques la vista paginada. Una vez terminado, comprobaré que el total de filas coincida con el contador oficial y que no existan celdas vacías en los tres campos solicitados. Si tienes experiencia en scraping con Python (requests, BeautifulSoup, Selenium) o herramientas similares y puedes generar el .xlsx sin alterar la estructura original, me será suficiente. Entrega prevista: archivo Excel ...
...accuracy and time-stamping are essential. Store everything in a structured database of your choice (PostgreSQL or MySQL are fine). The tables must let me query: • first-pull values • second-pull values • calculated deltas between the two Please build in a simple scheduler or CLI flag so I can trigger each scrape automatically via cron. Bet365 is heavily scripted, so headless-browser handling (Selenium, Playwright, or Puppeteer) plus proxy/anti-bot measures may be required; use whichever stack you’re comfortable with, provided it is well documented. Deliverables 1. Clean, runnable source code with setup instructions. 2. SQL schema and migration script. 3. README showing sample queries that compare initial vs. pre-game lines. 4. Brief note on ho...
...service must issue and validate JWT tokens for every request beyond the public health-check route. Token refresh, revocation, and a simple role model (“user” vs. “admin”) should be built in from the start. Flight data extraction I do not have official Iberia developer access, so we will need to pull the data ourselves. I’m open to whichever tooling you are most comfortable with — BeautifulSoup, Selenium, Scrapy, or a hybrid approach — as long as the final solution is headless, resilient to minor layout changes, and respectful of Iberia’s rate limits. Only flights that are bookable with Avios need to be captured; no hotel or car-rental data is required. Deliverables • Clean, modular Python code (FastAPI or Flask preferred,...