Seleniumcông việc
...tiêu – Lấy đầy đủ tên sản phẩm, giá hiện tại, giá gốc (nếu có), thuộc tính chính, tên shop, lượt bán, lượt thích, điểm rating và đường dẫn ảnh. – Crawler phải đi hết các trang trong danh mục, không bỏ sót, không trùng dữ liệu. – Kết quả xuất ra CSV / Excel; cột nào ra cột nấy, tiếng Việt không lỗi font. Yêu cầu kỹ thuật – Ưu tiên Python sử dụng Scrapy, Requests + BeautifulSoup hoặc Selenium nếu cần vượt qua cơ chế chống bot của Shopee. – Code sạch, chú thích ngắn gọn, dễ đọc; tôi cần toàn bộ source kèm hướng dẫn chạy trên Windows (có thể bằng file README). &nd...
tôi cần code 1 tool sử dụng selenium python để tự động hoá thao tác cho 4 tác vụ này trên tạo video từ ảnh ( image to video ) tạo video từ prompts ( text to video ) tạo video cố định nhân vật ( sudject reference ) tạo ảnh từ prompt ( image generation ) logic chi tết thảo luận thêm
Viết chương trình test một website bán hàng bằng wedrive selenium với python test một vài chức năng như đăng nhập, thêm vào giỏ hàng, xóa sửa sản phẩm . Mỗi chức năng có 3-4 test case có thể export kết quả pass- fail mỗi test case ra excel.
Mình đang cần tìm một bạn để dạy kèm cho mình sử dụng python để làm webscraping cho mình. Mình prefer các bạn có kinh nghiệm làm việc với thư viện selenium. (10 đô/ 1h)
viết tool trên sàn giao dịch chứng khoán theo yêu cầu của mình. Tool đơn giản và không phức tạp, quan trọng là sau khi viết tool xong dạy lại mình cách viết code.
Tự dộng đăng ký facebook bằng chrome selenium. Yêu cầu hỗ trợ đa tác vụ và sắp xếp vị trí các cửa sổ chrome gọn gàng trên màn hình
Gear Inc. đang tìm 1 bạn QA Automation Engineer có thể làm việc tại công ty ở Hà Nội trong vòng 2 tháng, giờ hành chính. Yêu cầu: - Có thể làm việc giờ hành chính tại công ty ở Hà Nội - Có ít nhất 3 năm kinh nghiệm về automation testing - Có kiến thức về HTML, CSS, JavaScrip, etc - Có kinh nghiệm về automation test trên web-based và mobile application, sử dụng automation tools như Selenium, Appium, Cucumber, etc Quyền lợi - Lương thỏa thuận cạnh tranh up to $2000/tháng cho công việc freelancer - Cơ hội hợp tác lâu dài, trở thành nhân viên chính thức của công ty (Remove...
Start-up của Singapore cần tuyển: on Rails developer: Phát triển hệ thống back end, hỗ trợ tính năng mới. 3+ năm kinh nghiệm phát triển backend trên Ruby on Rails (RoR) Chuyên môn trong RoR và các công nghệ web. Có khả năng giải quyết ác vấn đề phức tạp. Kinh nghiệm thử nghiệm TDD (RSpec, Capybara, Selenium) Phát triển và nâng cấp backend của App dựa trên cấu trúc tổng quan, tăng thêm tính năng cho App. Khả năng học hỏi nhanh và cao. Có khả năng giao tiếp Tiếng Anh developer: 2+ năm kinh nghiệm với JavaScript, phát triển front end phần mềm. Kỹ năng: JavaScript, HTML5, CSS3, ngôn ngữ phát triển web và ...
...successful candidate How it works: • You refer a candidate who needs interview support • Candidate confirms and takes the service • You receive $30 per successful referral • No limit on referrals – unlimited earning potential Technologies Covered: • Software Development: React, Node.js, Java, Python, .NET, Full Stack • Cloud & DevOps: AWS, Azure, GCP, Docker, Kubernetes • Testing: Automation, Selenium, Cypress, Manual Testing • Data Roles: SQL, Data Engineering, Data Analyst, AI • Engineering: Controls Engineering , Automation, Embedded, Electrical • Enterprise Tools: SAP, Salesforce, ServiceNow, Oracle Why partner with me: • Transparent process • Fast payout after successful referral • Long-term partne...
...post) unless properly bound to browser session Maintain real browser fingerprint consistency Preferred Technical Approach: Selenium / SeleniumBase integration Chrome DevTools Protocol (CDP) usage Network interception or replay In-browser JavaScript execution (fetch / XHR override) Proper handling of dynamic request signing if applicable Goal: To programmatically trigger the same successful API call that the browser makes, while preserving complete browser session authenticity. Deliverables: Fully working implementation Clean and optimized code Explanation of approach used Must be tested and verified against protected endpoint Required Skills: Python Selenium / SeleniumBase Chrome DevTools Protocol (CDP) Browser Network Debugging JavaScript (fetch / XHR ...
...unless properly bound to browser session Maintain real browser fingerprint consistency Preferred Technical Approach: Selenium / SeleniumBase integration Chrome DevTools Protocol (CDP) usage Network interception or replay In-browser JavaScript execution (fetch / XHR override) Proper handling of dynamic request signing if applicable Goal: To programmatically trigger the same successful API call that the browser makes, while preserving complete browser session authenticity. Deliverables: Fully working implementation Clean and optimized code Explanation of approach used Must be tested and verified against protected endpoint Required Skills: Python Selenium / SeleniumBase Chrome DevTools Protocol (CDP) Browser Network Debugging JavaScript (fetch / XHR ha...
...content in a clean CSV or Excel file with clear column headings; if you prefer a database export, let me know and we can adjust. • Include the finished script or notebook so I can rerun the extraction later. Accuracy and formatting matter more to me than sheer speed, so please allow time for basic validation before handing over the files. If you normally work with Python (BeautifulSoup, Scrapy, Selenium) or similar tooling, that’s perfect, but I’m open to alternative stacks as long as the output meets the same standard. When you reply, briefly outline: 1. The scraping approach and libraries you’d use 2. Any anti-blocking measures you apply for public sites 3. A realistic timeframe to capture, clean, and hand back the data I’m ready to start a...
Project Title: WhatsApp Transfer Verification & Auto Receipt System Project Descriptio...representative on WhatsApp. - Mark the transfer as approved. 5. Automatically open my company accounting system (web-based). 6. Create and submit a Receipt Voucher (Cash Receipt Entry) with the transfer details. Technical Requirements: - Python or Node.js - WhatsApp API integration (or automation if API not available) - OCR for reading transfer receipts (Arabic + English) - Web automation (Selenium / Playwright) - Secure handling of credentials - High accuracy verification logic Deliverables: - Full source code - Executable version (EXE if Windows) - Setup documentation - Training on how to use the system Goal: Automate transfer confirmation and receipt entry to save time and reduc...
I need a reliable browser-based bot that logs in to my Amazon A to Z account, continuously monitors upcoming shifts and instantly books the ones that match criteria I will later define (location, start time, hours). The tool must operate inside a standard web browser—headless Chrome or a Selenium/Puppeteer script is fine—as long as it books faster than manual clicking and survives Amazon’s usual page refreshes, captchas, and timers. I will grant the bot full access to my account solely for booking purposes, so handling login securely (2FA, encrypted credentials, cookie reuse) is essential. A simple configuration file or UI where I can tweak preferred warehouses, shift lengths, and daily booking windows would be helpful, but speed and reliability come first. Deliv...
...publicly available sites. The focus is simple yet crucial: for every company you find, capture the homepage URL and a working email address. (ask for details in the sheet ) A completely ethical approach is non-negotiable—no gated content, no third-party lists, and no automated harvesting that violates site terms. I’m happy for you to use tools you’re comfortable with (Python, Scrapy, BeautifulSoup, Selenium, Google Apps Script, etc.) as long as you respect and rate limits. Email addresses must appear in plain text within the sheet; please avoid hyperlinks or HTML encoding. Deliverables • A Google Sheet populated with data • A short note on your collection method (manual, scripted, hybrid) so I can replicate or update the data in the future. ...
...location de bateaux. Pour y parvenir, j’ai besoin d’un workflow automatisé reposant exclusivement sur Google Maps – c’est la source retenue – capable de collecter, dédupliquer puis nettoyer les données avant de les mettre en forme dans un CSV directement importable dans le CMS Wix. Les coordonnées devront être fournies en degrés décimaux. Livrables attendus • Un script réutilisable (Python + Selenium, Scrapy ou équivalent) qui interroge Google Maps, gère le rate-limit et documente chaque étape de traitement. • Le fichier CSV final contenant, pour chaque base nautique, les champs suivants : Nom, Adresse complète, Ville, Département, Régio...
...búsquedas y resultados Entregables Código fuente completo Documentación de instalación y uso Documentación de endpoints de la API Ejemplo de request/response (texto e imagen) Deploy listo en servidor/VPS o guía de deploy Importante El sistema debe ser escalable y mantenible, ya que se planea agregar más tiendas con el tiempo. Perfil buscado Experiencia comprobable en scraping (Playwright/Puppeteer/Selenium) Experiencia creando APIs REST Experiencia en visión computacional / reconocimiento de imágenes (deseable) Ideal si tiene ejemplos previos de scraping e-commerce o monitoreo de precios Para aplicar Por favor enviar: tecnologías que usarías enfoque para búsqueda por imagen (tipo Lens) ejemplos...
I am looking to build a custom AI agent that manages my end-to-end job application workflow. The goal is to create a semi-au...new matched jobs • Review drafted applications • Approve / reject submissions • Track applied jobs • View outreach sent • See recruiter responses • Track interview updates Clean UI preferred. Minimal but functional. ⸻ Technical Expectations Open to suggestions, but likely stack: • Python / Node backend • AI via OpenAI API or similar • Automation using browser automation (Playwright/Selenium) • Email integration (Gmail API) • LinkedIn automation (must comply with platform limits) • Cloud deployment (AWS/GCP/Azure) • Simple web/mobile-friendly dashboard (React / Flutter / etc.) ...
...Python crawler that pulls public content from Twitter, Instagram and LinkedIn, covering text, image and video posts for any handle I feed it. Here’s the flow I have in mind. The script collects the raw post data (caption, hashtags, basic engagement numbers and, where accessible, image/video URLs) through whichever mix of libraries makes sense—Tweepy or Twitter API v2 for Twitter, Instaloader or Selenium for Instagram, and the official or unofficial LinkedIn API for LinkedIn. After normalising everything into a common JSON schema, the crawler should pass that dataset to an LLM endpoint (OpenAI or similar) and receive back a concise, structured report that includes: • Brand sentiment (positive / neutral / negative trends) • Key thematic buckets the bran...
Project Title Custom Lead Generation & Email Scraper Tool (Google, Yellow Pages, & B2B Directories) Project Description I am looking for an experienced developer to build a robust, high...Email Address (Must have a validation check to avoid "dead" emails) Phone Number (Optional but preferred) LinkedIn Profile URL (Optional but preferred) Export Functionality: Capability to export data into CSV or Excel format. Anti-Blocking Measures: Use of rotating proxies or delays to ensure the scraper isn't blocked by Google or directories. Technical Requirements: Preference for Python (Selenium, Scrapy, or BeautifulSoup) or a dedicated desktop application. User-friendly interface (even a simple CLI is fine, but a GUI is a plus). Fast processing speed with the abil...
Project Title Custom Lead Generation & Email Scraper Tool (Google, Yellow Pages, & B2B Directories) Project Description I am looking for an experienced developer to build a robust, high...Email Address (Must have a validation check to avoid "dead" emails) Phone Number (Optional but preferred) LinkedIn Profile URL (Optional but preferred) Export Functionality: Capability to export data into CSV or Excel format. Anti-Blocking Measures: Use of rotating proxies or delays to ensure the scraper isn't blocked by Google or directories. Technical Requirements: Preference for Python (Selenium, Scrapy, or BeautifulSoup) or a dedicated desktop application. User-friendly interface (even a simple CLI is fine, but a GUI is a plus). Fast processing speed with the abil...
My team has a fully-built web application and I need a fresh set of expert eyes to validate every line of code. The goal is to run a complete test cycle—functional, performance, and security—so we can ship with confidence. You are free to recommend and use whichever framework or tool best fits the job; I am open to industry standards such as Selenium, Playwright, JMeter, OWASP ZAP, or anything you feel will surface issues quickly and reliably. What matters most is deep coverage of the entire source code, clear defect reporting, and actionable suggestions for hardening and optimisation. I will provide the full repository, environment details, and access credentials once we agree on the engagement. In return, I expect: • A concise test plan outlining scope, assump...
I already have a curated list of LinkedIn profile URLs and need the key networking details moved into a single Google Sheet. For every profile, capture each person’s stated interests and list the five types ...(as it appears) • Interests (comma-separated) • Type 1 through Type 5 (verbatim wording) • Category tag (Industry experts / Potential clients / Collaborators) Accuracy of the text you pull is more important than speed, but I do expect the work to comply with LinkedIn’s terms and avoid triggering any scraping limits. If you prefer to work manually, that’s fine; if you script with Python, Selenium, or similar, just make sure the final output lands cleanly in my shared Google Sheet. I’ll review a small sample first, then green-light...
...it detects any difference in either the visible text or the images. The page updates unpredictably, so true real-time tracking is essential; a one-minute polling interval is already too slow for my use-case. A lightweight approach that respects the site’s bandwidth and avoids triggering blocks or captchas will be valued. I am open to whatever stack you favour—Python with BeautifulSoup or Selenium, Node.js with Puppeteer, or a compiled solution—so long as it is stable on a Windows environment and easy for me to tweak the target URL later. Notification method is flexible: an email is fine, but if you have a smarter suggestion (desktop toast, webhook, etc.) that can be set up quickly, I’m all ears. Please deliver: • The complete, well-commente...
I’m spinning up additional quality coverage for an active web and mobile project and need a seasoned tester who can switch easily between writing code-driven checks and running thoughtful manual sessions. Your day-to-day will include drafting concise test plans, building and maintaining Selenium-, Cypress- or Playwright-based suites, and verifying APIs with Postman or a similar tool. When something breaks, I’ll count on you to log reproducible defects and follow them through to closure, collaborating directly with developers in our Agile board so each sprint ships clean. Because delivery timelines are tight, I’m specifically after someone who already has strong automation foundations, understands functional, regression and end-to-end coverage, and can work indepen...
Title: Contract QA Automation Engineer – Onsite (3–6 Months) Project Overview: We are hiring QA Automation professionals for a short-term contract (3–6 months) to support active development projects. Location (Onsite/WFO): Udaipur | Jaipur | Bangalore | Bhopal Skills Required: Proven experience in automation testing Selenium / Cypress / Playwright or similar tools API testing knowledge Strong problem-solving skills Contract Details: Duration: 3–6 months Full-time engagement Onsite only (no remote option) Who Should Apply: Freelancers or contract professionals available for onsite work Immediate to short notice joiners preferred
...using Playwright, one of the fastest-growing and most powerful automation frameworks in the market. This comprehensive course is designed to take you from complete beginner to advanced automation engineer through hands-on projects, real-world scenarios, and framework development. Playwright Full Course - Complete Table of Contents Beginner Level • Introduction to Playwright • Why Playwright vs Selenium and Cypress • Supported Browsers and Languages • Environment Setup (Node.js, IDE, Installation) • Project Structure and Test Runner • Understanding Locators and Selectors • Basic Actions: Click, Fill, Type, Select • Assertions using Expect API • Handling Waits and Timeouts • Debugging with Inspector and Codegen • Running an...
...pages in real time or on a frequent schedule. • Apply NLP or other classification techniques to decide whether a posting is truly AI-related, then tag it by sub-domain (e.g. vision, NLP, MLOps, prompt-engineering). • Deliver concise, deduplicated listings to me through an in-app notification feed—no email or SMS required. For the deployment side I’m open to Python (Scrapy, BeautifulSoup, Selenium), Node, or any stack you are comfortable with so long as it is containerised and can run unattended on a small cloud instance. A lightweight web interface or Electron desktop app for the notification feed is ideal; you can suggest an alternative if it achieves the same user experience. Acceptance criteria 1. Agent successfully scrapes at least three live sou...
...captured each record logged beside the data. • Consistency matters: please apply uniform naming conventions (e.g., “FY2023 Gross Profit” instead of varying labels) and check subtotals or totals to be sure everything reconciles. • I’m flexible on the final file type—CSV, Excel, or Google Sheets all work—so let me know which you prefer or suggest. • If you automate with Python, BeautifulSoup, Selenium, or a comparable tool, great; just include the script so the process can be rerun later. A quick README explaining any inputs or environment setup is enough for me to replicate it. • Accuracy is non-negotiable. I will spot-check figures against the original web pages, so double-check before submitting. Once you deliver the dat...
I need end-to-end testing for my web application. The project code is already written, so you will be focusing solely on testing. Requirements: - Thoroughly test on Chrome - Ensure all workflows function as intended - Identify and document any bugs or issues Ideal Skills: - Experience with end-to-end testing tools (e.g., Selenium, Cypress) - Strong attention to detail - Familiarity with web applications and browser testing Please provide a brief overview of your testing experience and any relevant tools you plan to use.
I need a reliable solution that can pull data from LinkedIn and insert it straight into a database I specify. The core requirement is the automated transfer—once the tool fini...strings and add new profile URLs without touching the core logic Acceptance criteria • I point the tool at 10 test profiles; all mapped fields appear in the database exactly once with no missing characters • A command-line log or dashboard shows progress and any errors in real time • You hand over the source code, a README with setup steps, and a sample SQL schema Experience with Python (Selenium, Playwright or similar), LinkedIn APIs, and database integration will be highly valued. Let me know your preferred stack, any relevant past projects, and the estimated timeframe to del...
I need a straightforward Python script that signs in to a Hotmail / account with Selenium (or another reliable browser-automation library)
I need end-to-end testing for my web application. The project code is already written, so you will be focusing solely on testing. Requirements: - Thoroughly test on Chrome - Ensure all workflows function as intended - Identify and document any bugs or issues Ideal Skills: - Experience with end-to-end testing tools (e.g., Selenium, Cypress) - Strong attention to detail - Familiarity with web applications and browser testing Please provide a brief overview of your testing experience and any relevant tools you plan to use.
I have a public-facing website that I need scraped end-to-end. The site is open (no login), but the content is split across multiple pages, so your script will have to detect and follow pagination automatically. Here is exactly what I expect: • A clean, well-commented Python script (requests/BeautifulSoup, Scrapy, or Selenium—your choice) that visits every page, captures the required fields, and writes them to a neatly structured CSV. • The final CSV containing all rows pulled from the site. • A short README that tells me how to run the script and change the target URL or output path if needed. Code quality matters to me: no hard-coded absolute paths, clear variable names, and graceful error handling so the run doesn’t stop if a single page fa...
I...framework into a CI pipeline. • Walk-throughs of flaky-test triage, mocking external dependencies, and debugging failures that only show up in complex environments. • Short take-home exercises or sample repositories that reinforce each lesson, plus code reviews so I know I’m applying the patterns correctly. Although my main interest is integration testing, I’m flexible on specific tooling: Selenium, NUnit, SpecFlow or whichever stack you feel showcases best practices in C# automation. The important part is understanding why a tool is chosen and how to extend or swap it later. Please outline your preferred tools, how you normally structure a learning path, and roughly how many hours you expect we’ll need to reach a self-sufficient framework and...
...support release cycles. • Maintain test documentation and contribute to continuous improvement of QA processes. Required Skills & Qualifications Core QA Skills • 3+ years of experience in software testing, preferably in cybersecurity or networking domains. • Strong understanding of QA methodologies: black-box, white-box, and grey-box testing. • Experience with test automation frameworks (e.g., Selenium, PyTest, Postman). • Familiarity with reverse proxy tools (NGINX, HAProxy, Envoy) and network traffic analysis. • Hands-on experience with Linux environments and shell scripting. • Graduate from a Top 50 engineering college in India as per NIRF 2025 ranking. Cloud & Security Awareness • Testing in cloud platforms (AWS, Azure, GCP) ...
...implement a robust test automation architecture with Maven/Gradle and CI/CD pipelines Page Object Model (POM) or similar design pattern reporting (Allure/Extent Reports or similar) reusable and maintainable test scripts documentation for setup and usage Required Skills: experience in Playwright with Java understanding of Selenium/automation concepts with TestNG/JUnit in CI/CD integration (Jenkins/GitHub Actions, etc.) of Git version control in framework design and best practices Nice to Have: automation experience knowledge testing exposure Duration: Short-term (with possible extension) Share your previous Playwright (Java) project
...CSV should include original variables like organization name, state and zip even though that data was not used in the scraper. The script must perform the following steps for each URL in the input list: 1. Input: Read a list of URLs from a provided CSV file (single column of URLs). 2. Navigation/Rendering: Visit the URL (handling redirects is essential). The use of a headless browser (like Selenium/Puppeteer) or an advanced HTTP library is preferred, as some websites may load the footer content dynamically via JavaScript. 3. Targeted Scanning: Scan the HTML source code of all pages found in the sitemap, specifically looking for the presence of a specific link. 4. Output Logic: - If the link is found, record the identified vendor. - If no vendor is explicitly identified, ...
I need the entire history of a specific Facebook Group captured—every post along with all associated comments. I’m ...with working links to the images and videos placed in clearly named folders. I don't want folders or links. Just one huge continuous page that has everything. This is for a court case and I have to give this to the other side. I want them to have to scroll through however many hundred pages there are. Just as if they were actually on FB. Please outline: • your scraping approach (Python + Selenium, Go, node-puppeteer, etc.), • how you’ll handle media downloads and folder structure, • estimated turnaround time. I’ll review a short sample export before we proceed with the full run to confirm the layout meets my ...
...need a Selenium-based solution that runs reliably on Windows and opens Google Chrome to simulate human visits to LinkedIn (and occasionally other) profile URLs listed in a Google Sheet. For each URL the program should: • Pull the next unused link from the sheet • Load the page in Chrome, wait a random time between 20 seconds and 3 minutes • Apply truly randomized scrolling patterns while the profile is open so behaviour looks organic • Fire a webhook the moment the visit completes, passing back any ID or payload I define so our CRM reflects the touch instantly Configuration items such as Google Sheet ID, webhook endpoint, minimum/maximum dwell time, and daily visit caps should live in a simple file I can edit without touching code. A short README on ...
...the comment text, number of comments, likes, reposts/shares, the post date and any other readily available metadata (author handle, follower count, post URL, media links, etc.). Accuracy is critical because the data will feed a trend-analysis dashboard later. Please build the workflow in a way that respects rate limits and login requirements: if you intend to use official APIs, private APIs, Selenium, Scrapy, Playwright, or headless browsers, spell that out so I know how sustainable the solution will be. The final hand-off should include: • A clean, well-commented reusable script (Python preferred) • A short README explaining environment setup, keyword input format and how to extend to new regions • The full export in CSV so I can validate before sign-off I...
I need a r...I can decide the exact days and times each role goes out • one-click “auto post” that instantly publishes a job when I hit the button The posting frequency isn’t fixed; some days I may blast several openings, other weeks none at all, so the scheduler has to respect whatever plan I set. Use whichever method makes the process most stable—headless browser automation (Puppeteer, Playwright, Selenium) or direct API calls if Placementindia offers them. The final package must run on a standard VPS, expose clear logs/errors, and allow easy editing of job templates. Please share links or short videos of similar bots you’ve built so I can gauge robustness. Once I can log in, queue a job, and watch it appear live without intervention, I&rs...
...Primitiva: €1.00 El Gordo: €1.50 Dynamic AI suggestions: recommend bet quantity based on backtesting, current jackpot and balance, using Monte Carlo for EV (prioritize positive EV scenarios). 2.5. Login and Bet Execution Automation Never store login/password (not even encrypted). Prompt user for credentials (email/NIF/NIE + password) every login, using secure masked fields. Login process: Selenium (headless Chrome/Firefox), access , fill form, submit, verify success (post-login elements), handle sessions/cookies. Bet execution: navigate to lottery page, generate AI numbers, add exact user-requested bets to cart (batches for high volumes), support rules of each lottery. Finalization: show full summary (lottery, bets, cost, numbers); require
I need a small, Windows-friendly Python script that will open a real browser with Selenium and wipe large batches of content from my X (Twitter), Facebook, and Instagram accounts. Because my X account sits on the free API tier I keep running into 403 errors, so this project must rely solely on browser automation—no official APIs or paid third-party tools. Here’s what I’m after: the script launches from the command prompt, asks for (or reads from a .env) my login credentials, signs in, and then iterates through all visible posts, tweets, and reels, deleting each one until none remain or until it hits an optional stop condition such as a date or a post count I can set. A simple console printout like “Deleted tweet #42” is enough for logging; I don’...
IM TYRING TO RUN THE ATTACHED JPNY SCRIPT TO GET INFO FROM A WEBSITE BUT I CANT UNDERSTAND IT DOESN'T WORK. I NEED THIS SCRIPT TO BE FIX + PAGINATION TO FETCH AROUND 2400 RECORDS FOR YELLOWPAGES I ONLY USE JUPYTER
...market research. The job centers on extracting selected data points from public web pages, transforming them into a clean, structured format, and making them available for analysis every 24 hours. Here’s what I need you to handle from end to end: • Source acquisition – fetch HTML from the URLs I provide, even when content is hidden behind JavaScript (a headless browser such as Playwright or Selenium is fine). • Parsing & cleansing – pull the specific fields I’ll list (product name, price, SKU, availability, and a time-stamp), remove duplicates, and standardize values. • Storage & delivery – load the daily output into my PostgreSQL instance; if you prefer Parquet or plain CSV that’s acceptable as long as it’s a...
...information should be organised into a clean CSV file—one row per page—with columns for page URL, full body text, image file names, and link destinations. Please download the images themselves as well and bundle them in a separate folder (a simple ZIP is fine); the CSV should reference the exact filenames so everything lines up. I’m happy for you to use Python with BeautifulSoup, Scrapy, Selenium or whichever stack you prefer, as long as the final output meets these acceptance criteria: • Complete CSV containing text, image names, and link URLs for each page • All images successfully downloaded and accessible via the filenames listed in the CSV • No duplicates or missing pages from the target site * Images need to be sorted for each l...
I am looking for a Python developer to create a simple and focused scraper script for Facebook Mar...• A file containing all product URLs for that seller • File format: TXT or CSV • Handle infinite scrolling to load all products Technical Requirements: • Python • Selenium or Playwright • Experience with dynamic websites • Clean, runnable, and well-structured code Important Notes: • No filters required (no country, city, or keywords) • No data is needed other than product links only • Manual login can be used if required Budget: Open — to be discussed based on experience and quality When Applying, Please Include: • Any previous experience with Facebook Marketplace scraping • The tool you plan to use (...
I have a data-analysis pipeline that relies on a steady flow o... • Payload: high-resolution image files plus a CSV/JSON map linking each file to product ID, title, price, and category text that you extract during the same run. • Scale: thousands of products per crawl; a resumable approach is essential so partial failures don’t force a full restart. • Frequency: I’ll trigger the crawl weekly, so reusable code is a must. I’m happy with Python—Scrapy, Selenium, Playwright, or a headless solution of your choice—as long as it respects the site’s anti-bot measures and keeps requests polite. Please include a brief outline of how you’ll handle pagination, lazy-loaded images, and rate limiting. Let me know your proposed stac...
...mandatory declarations and any digital signature steps must all be handled by the system before it attempts the final submission. A short, clear dashboard that shows “ready”, “errors found” or “submitted” status for each tender would be ideal so I can intervene only when something is missing. Deliverables I must see to accept the project: • Source code and install guide (preferably Python with Selenium / Playwright or similar RPA layer, but I’m open if you can justify another stack) • A configuration file where I can add new government portals without touching core code • Automated validation rules that stop a submission if any mandatory field or attachment is missing • A post-submission PDF/CSV log summarisin...