Crawler regexpráce

Filtrovat

Moje poslední hledání
Třídit podle:
Rozpočet
pro
pro
pro
Typ
Dovednosti
Jazyky
    Stav zakázky
    2,000 crawler regex zakázek nalezeno, ceny v EUR

    To complete a part of a research study, I'm looking for a python developer to build a fast and efficient web crawler model to retrieve 1) "financials" and 2) "news articles" related to all NYSE stocks in the past 20 years. It's an expanded project and the candidates need to determine where to access the financial market news as well and financial metrics. Also, due to the high volume of the data that will be scraped, the contractors need to have the equipment and high speed connection, necessary for the job.

    €165 (Avg Bid)
    €165 Průměr. nabídka
    8 nabídky

    Main#1) Obtain your corpus of documents for the semester. To do so, come up with 10 neutral (ie no controversy) queries (for example: Who was the 16th President?) that you will submit to your search engine. You are to then download the first 20 (non-controversial) webpage responses that the ...controversy) queries (for example: Who was the 16th President?) that you will submit to your search engine. You are to then download the first 20 (non-controversial) webpage responses that the search engine returns with, for each of the 10 queries. There will be a total of 200 html files. (We will be discussing shortly how to process these using the Java Regex package. You may NOT use 3rd party code. You MUST write your own. You do not need regex necessarily but it does provide much mo...

    €28 (Avg Bid)
    €28 Průměr. nabídka
    3 nabídky
    Crawl / Scrape Job Ukončeno left

    In the first step, the car models and their years of manufacture (from 2005) are to be crawled from a website and written to a CSV file. In the second step, the models will be searched for on another website, the detail page will be called up and the images downloaded. The images should be provided in a tar archive (e.g. AWS S3). The exact websites and 3 e...crawled from a website and written to a CSV file. In the second step, the models will be searched for on another website, the detail page will be called up and the images downloaded. The images should be provided in a tar archive (e.g. AWS S3). The exact websites and 3 examples are included in the attached PDF document. The whole process should be automated. Summarized: - Write 2 Website Crawler - Provide tar archive of ...

    €74 (Avg Bid)
    €74 Průměr. nabídka
    13 nabídky

    ...Dell, Urban Outfitters, AdBlock, and VitaminShoppe. Technical SEO Consultants are key players on every client team, serving as our go-to experts on tricky technical SEO questions that come up in our work with our largely startup, ecommerce, and SME clientele. Here's what your typical week might look like: Monday: conduct a thorough technical audit for a new startup client using your favorite crawler and other tools Tuesday: recommend and implement a solution for a client with 4,000 tag archive pages on their Wordpress blog Wednesday: hold a training for our consultant team explaining how to prioritize site speed recommendations according to the balance between impact and implementation difficulty Thursday: setup a reverse proxy to rewrite URLs so a client can host the...

    €25 / hr (Avg Bid)
    €25 / hr Průměr. nabídka
    39 nabídky
    Crawler data Ukončeno left

    We are looking for a developer who is experienced in building a web crawler. ( removed by admins )

    €435 (Avg Bid)
    €435 Průměr. nabídka
    11 nabídky

    Looking for a SW developer who has built web crawlers to scan websites of specific text strings and specific graphic files ( all web formats ) to check for illegal logo usage to allow me to offer a logo/trademark checking/validation service to website owners. I am flexible on the code and product base and I am open to use of Open Source products

    €1011 (Avg Bid)
    €1011 Průměr. nabídka
    16 nabídky

    ...A word is considered to be one or more characters in length separated by white space (space character, tab, new line character)or a comma, period, exclamation pointor question mark. Quotes may appear in a word but these mustbe ignored and not included as part of the apostrophe character may also appear in a word. You are provided with a regex that will filter these characters out during the file reading not change the regex as doing so will possibly result in your results not matching the correct results. BookWord will be used to holdthe characters for each unique word as a string along with acount which will contain the number of occurrencesof that wordin the novel. The main method will create an ArrayList of BookWord to hold the Words for the novel. The starting c...

    €20 (Avg Bid)
    Neodkladné
    €20 Průměr. nabídka
    5 nabídky

    ...read the attached 1) Word document for instruction 2) Excel sheet - all 4 pages and strictly follow the instruction and send me your sample work then I will not respond to your messages. Automatic bidding will be automatically rejected. Your first message MUST accompany your sample work in the attached Excel sheet. Otherwise, I won't even read your message. Using your own web crawler script or by any other means you will populate attached Excel sheets 1) 10,000 Professors 2) 50,000 Graduate Students 3) 20,000 Employees (who are working for high tech businesses in high-tech job). A total of 80K (80,000) emails You will do this in 8 milestones for equal payment of 8 x $50.00 = $400.00 I will pay to Freelancer in advance for each milestone. Now, read...

    €408 (Avg Bid)
    €408 Průměr. nabídka
    42 nabídky

    I would like to let program a Python Program that search and discover on all Domains and Subdomains on a specific IP address all Entry's, Files and Directories that contain my name and output me the related URL too. I would like to talk with you about my Project Suggestion.

    €172 (Avg Bid)
    €172 Průměr. nabídka
    5 nabídky

    We want to scrape real estate listings of 3 different websites. More websites will be added later. Ideally we can add more websites by configuration (ideally we can configure the regex for each variable). We will provide a briefing with all values marked on the respective listings pages. The goal for the tool is to scrape the newest listings so we can check them and if needed, post them on our own platform (of course with a reference to the source) instead of copy-paste them by hand. The scraper should have the following behavior: - 1st we scrape the index/result pages to find newly added listings - 2nd we visit and scrape individual listings The scraper should not scrape more than 5 listings per hour per website to prevent unnecessary load and the scraper from being banned. Sc...

    €540 (Avg Bid)
    €540 Průměr. nabídka
    19 nabídky

    I need to have a crawler php that will copy some information from one url to a wordpress database. DO NOT QUOTE WITHOUT READ THE FULL WORD

    €395 (Avg Bid)
    €395 Průměr. nabídka
    47 nabídky

    We need a single developer or a team of developers to write around 5000 lines of regular expressions. No other programming language is needed, all you need to write is pure regex, nevertheless, you still need to be clever as the regex requires extended thinking and advanced knowledge. Need someone with devotion that can work long hours daily because we need this task done asap.

    €1335 (Avg Bid)
    €1335 Průměr. nabídka
    12 nabídky

    ...to fetch stream http://serverip:port/ --------------------------------------------------------------------------------------- i am using a script for live youtube chanel i have uploaded the script so you can see what i am trying to acheive this is how i am using it 1 ) Regex for every part of the script 2 ) GET methods to make sure they are set & if not set them to null 3 ) Added a switch statement for quality selection 4 ) Removed some unneeded code with new regex 5 ) Made it echo out an error depending on if it failed to --- Grab videos --- Didnt find a video id -- -Didnt find the m3u in page To use this script you will need to upload it to the www directory of your server that you wish to restream from If your using XtreamCodes you nee...

    €116 (Avg Bid)
    €116 Průměr. nabídka
    3 nabídky

    ...wordpress based website. We want to enhance the content of our website with a list of events worldwide. As data source we want to use 80 websites from AHKs world wide: They are all in a simliar format and those websites are based in Typo3. So we need to put all of those 80 AHK websites in our crawler software and crawl them daily for keywords like "Training Programme" , "Geschäftsreise", "Konferenz", so the keywords which are in the search bar under the category. After we have that info, we want to show them on our website with the perspective country/region. But that's easy because each website we use as a source just displays events from their country. Please give me an estimate on the

    €497 (Avg Bid)
    €497 Průměr. nabídka
    43 nabídky
    page crawler Ukončeno left

    The goal ist to send the all private landlords from this free search result lists a message through the plattform's form. Detailed Description:

    €446 (Avg Bid)
    €446 Průměr. nabídka
    31 nabídky

    What we're looking for: a developer to create a crawler to obtain data from a website and save it in a MySQL DB from a Wordpress site. Versioning will occur through commits using BitBucket. In your proposal, please share a brief summary of your experience and tell us about a recent full stack web development project you worked on.

    €44 (Avg Bid)
    €44 Průměr. nabídka
    12 nabídky

    1. Get all the products details information from website 2. Export all the products information to XSLS file

    €21 (Avg Bid)
    €21 Průměr. nabídka
    22 nabídky

    I need an experienced C developer with experience of projects using epoll to build a web crawler capable of making 10,000 concurrent connections. See the C10K problem for more details of what is required to make this work. I have decided on an epoll based architecture on a linux platform. See wrk on github for the kind of underlying functionality this project needs and the performance I am looking for.

    €690 (Avg Bid)
    €690 Průměr. nabídka
    4 nabídky
    Video URL Parser Ukončeno left

    I need you develop video URL parser in java. Parse a video URL source from any link provided. Knowledge of webscraping or web crawler is helpful but not required. You can use any open-source java library to create video URL parser. Write "books" in proposal so I know you have read project description.

    €37 (Avg Bid)
    €37 Průměr. nabídka
    5 nabídky

    we are market research agency. we need to grab item desctiption, price, action, regular price and some other item data from webstore. Approx 25,000 lines per store. scan will be performed weekly or twice per week. output file: *.csv (or excel or similar).

    €131 (Avg Bid)
    €131 Průměr. nabídka
    20 nabídky

    Hello, I am trying to extract the full name,first name and last name from an email. The data comes in as follows: | Name | Diana Yu | What gets extracted from this is: fullname: Name | Diana Yu | firstN: Name lastN: | Diana Yu | The code is: code: import string import re weblead = input_data['webleads_body'] value = ('n', 26) regex= (".*(Name).*") name = [(0) for l in value for m in [(l)] if m] fullname = name[0] fullname = (' ', 1)[1] fullname = () firstN = (' ', 1)[0] firstN = () lastN = (' ', 1)[1] lastN = ()

    €2 / hr (Avg Bid)
    €2 / hr Průměr. nabídka
    4 nabídky

    Hi, I need someone to develop me a fast and powerful web crawler to extract data from Google SERP. Important key features; 1. Must have a user interface protected by user and password where I can manage my campaigns. 2. Must have rotating proxies support 3. Multi threading support I want a dashboard, where I can create/edit/delete my campaigns. For every campaign, I will set a google query (ie: intitle:"fashion" "blog") and decide how many results I want to extract. When I will click on the "Start Crawler" button, I want to see the results in real time.. so If I decided to crawl the first 100 000 results for this query, if the crawler have extracted 1000 results, it should show 1000 results extracted out of 100 000. Once the extracti...

    €553 (Avg Bid)
    €553 Průměr. nabídka
    14 nabídky

    Built an open source crawler to scrape e-commerce sites by Category (prices & product catalog, key elements of the commence site). Focused search criterion or a topic. Design and Implement 1. Provide users login with Facebook and Gmail email account and p/w 2. Provide users to search and gather content information by categories 2. Create a copy of all the visited pages for further processing by admin. 3. Use a web crawler to extract data on real time changes & competitors trends. 4. Collect user info, data, marketing data, campaigns and use this data to make more effective and marketing decisions. Validate hyperlinks & HTML codes. 5. Make and Create Social Media pages. 6. Website should be mobile friendly. Price of the project should include 1 year maintenance...

    €231 (Avg Bid)
    Doporučené
    €231 Průměr. nabídka
    28 nabídky

    Quote to Supply and Install Google recaptcha to prevent crawler bots contaminating the database. oscommerce

    €67 (Avg Bid)
    €67 Průměr. nabídka
    15 nabídky

    Regex Chrome Extension Autofill Brief to integrate chrome regex autofill files into our existing extension. Qualifications needed Java script backend Experience working with Must be willing to sign an nda Functions Must have regex fields for the following First name Last name Full name Email Dob Address line 1 Address line 2 City County / state Country Postcode Phone Checkbox Name on card Card number Expiry month Expiry year Cvv You should be able to easily add new element ID to the file e.g for sites the autofill does not currently work on Additional information To apply please message your portfolio / github and any other supporting information that will aid your application. No recruiters

    €188 (Avg Bid)
    €188 Průměr. nabídka
    5 nabídky

    ...each file in the list one at a time 3. Match/Search regex 4. Extract 2 strings of data (needs some magic) 5. Write str(message) to file. File name is str(email) 6. If file str(mail) exists, append a '-1' to name. If that exists, auto-increment. 7. Delete the original file from folder_path 7. Continue. Next file in list. I have a folder full of files: folder_path = '/tmp/test/' Use globs to create a list of all the files. An example file is attached. Is is an HTML file being saved as HEX/ASCII I think. You decide how to open the file. You may want to import codecs: f = ('', 'r', 'utf-8') Read each file in the list (one at a time - for loop) If the file does NOT contain the regex 'Subject: Confirm ...

    €112 (Avg Bid)
    €112 Průměr. nabídka
    19 nabídky
    Regex help Ukončeno left

    Replace Stop loss with 3rd entry, remove 3rd entry Also remove % from Entry1/2 - or even remove 20%, 25% across entire text if easier Turn this: ⚡️⚡️ #BTC/USD ⚡️⚡️ Signal Type: Regular (Long) Leverage: 10.0X Entry Targets: 1) 10380 - 20.0% 2) 10239.5 - 25.0% 3) 10075 - 55.0% Take-Profit Targets: 1) 10454 - 20.0% 2) 10560.5 - 20.0% 3) 10719 - 20.0% 4) 10843 - 20.0% 5) 10980 - 20.0% Stop Targets: 1) 9801 Into this: ⚡️⚡️ #BTC/USD ⚡️⚡️ Signal Type: Regular (Long) Leverage: 10.0X Entry Targets: 1) 10380 2) 10239.5 Take-Profit Targets: 1) 10454 - 20.0% 2) 10560.5 - 20.0% 3) 10719 - 20.0% 4) 10843 - 20.0% 5) 10980 - 20.0% Stop Targets: 1) 10075

    €12 (Avg Bid)
    €12 Průměr. nabídka
    1 nabídky

    Budget: $400.00 If you do not follow the instruction below then I will not respond to your messages. Automatic bidding will be automatically rejected. Your first message MUST accompany your sample work in the attached Excel sheet. Otherwise, I won't even read your message. Using your own web crawler script or by any other means you will populate attached Excel sheets 1) 10K Professors 2) 50K Graduate Students 3) 20K Employees (who are working for high tech businesses in high-tech job). A total of 80K (80,000) emails Last sheet, in the attached Excel sheet shows the universities you will target and the majors, department, you will target. You will get paid $50.00 for every 10K email collection (for each milestone). You will have 8 mile stones. All emails must be non-duplic...

    €661 (Avg Bid)
    €661 Průměr. nabídka
    42 nabídky
    Web Crawler Ukončeno left

    Hi, My name is Eduardo, I'm from Brail, I don't speak English, that's why I'm using Google Translator. I am looking for a professional who can build a web crawler to monitor more than 50 thousand news sites (portals and blogs in Brazil) daily and store on my server only news that have certain keywords, I need to store the news title, news date and the news text. I currently have servers on Amazon (AWS), I would like to keep all infrastructure on AWS. I would like an idea of how many servers I would need on AWS to support this demand. I also need to know if for each of the monitored sites, I need to manually "map" so that I can extract the date, title and text of the news, or if that would be automatic. And finally, an idea of the value of this JOB...

    €4289 (Avg Bid)
    €4289 Průměr. nabídka
    3 nabídky

    I want a Python web crawler should have all IP v4 addresses from iprange = 0.0.0.0 to search. Internal IP addresses can be ignored. If an IP address is assigned to a Darknet URL (.onion), the following data should be written to an xlsx file (one xlsx file can save all results). Darknet URL IP address Time stamp with time zone information These three points are very important. If the IP address is assigned to another website (not .onion), the result is ignored. Likewise, all IP addresses that are assigned to Internet users. The computing process can be done via CPU and graphics card (CUDA - if possible): If the program is stopped, the last used IP address is saved and the next time the next

    €553 (Avg Bid)
    €553 Průměr. nabídka
    6 nabídky

    Hello i see you completed a project for this one : freelancer(dot)com/projects/php/Buatkan-Plugin-Wordpress-Auto-Grabber/details like this one : mangabooth(dot)com/product/manga-lhscan-net-crawler/

    €237 (Avg Bid)
    €237 Průměr. nabídka
    1 nabídky

    Your task is to develop a page crawler in selenium, which fetches all SERP details of G-Search. The crawler fetches: - all results from the page in a structured way like the JSON here: - crawler is able to fetch up to a defined result e.g. 120 so basically you need to - collect all results - collect the target URLs - use PageObject pattern and other recommended from seleniumHQ Your input of your API function will be a valid G-search URL The output of your function will be a java-model-class containing all details. What is NOT needed: - a UI (not required) - json output (not required) - a service architecture (like spring or JEE) - persistency Deliver simply a function, which executes selenium and collects the given details To execute the test, create a JUnit-run...

    €155 (Avg Bid)
    Doporučené
    €155 Průměr. nabídka
    10 nabídky

    I have Wordpress website with Woocommerce. We want to add an extra registration field at which you have to enter your VAT number and preferably we want a system that can check if the VAT is legit (trough a regex?). So Javascript/PHP experienced person with knowledge of Wordpress database.

    €131 (Avg Bid)
    €131 Průměr. nabídka
    34 nabídky

    Generell geht es um einen Amazon Scraper/Crawler , welcher bestimmte Produktdaten crawlt und diese in eine Datenbank einpflegt. Dies soll mit Hilfe der Amazon MWS API geschehen. Das Programm wird auf einem externen Server laufen und soll selbständig automatisiert einmal am Tag laufen und die Datenbank aktualisieren. Es geht um Datenmengen bis ca. 500.000. Eine genaue Beschreibung der Begriffe, Anforderungen und des Vorgehens ist vorhanden. Vielen Dank

    €37 / hr (Avg Bid)
    €37 / hr Průměr. nabídka
    5 nabídky

    I want to build simple web crawler using PHP Phalcon. Using PHP Phalcon, crawl up to 5 pages of this website (). It should output the following: - Number of pages crawled - Number of a unique images - Number of unique internal links - Number of unique external links - Avg page load - Avg word count - Avg Title length - Table to display each page crawled and it's status code

    PHP
    €32 (Avg Bid)
    €32 Průměr. nabídka
    2 nabídky

    create proper settings for OCTAPARSE to collect basic seller contact data from ETSY WEBSITE. OCTAPARSE IS A WEB CRAWLER THAT I NEED TO CREATE CUSTOM SETTINGS TO COLLECT SELLER INFORMATION FROM THE ETSY WEBSITE. THE KEY FEATURE IS THAT THE WEBCRAWLER SHOULD BE ABLE TO SEARCH BASED ON KEY WORDS , OR SOMETHING RANDOM, WE ARE LOOKING FOR A GOOD DATA COLLECTION TOOL 5 DATA SETS NEED TO BE COLLECTED 1) SELLER STORE TITLE NAME 2) SELLER CONTACT - EMAIL OR PHONE 3) SELLER WEBPAGE- URL LINK 4) SELLER NAME 5) LOCATION

    €116 (Avg Bid)
    €116 Průměr. nabídka
    14 nabídky

    I want to build simple web crawler using PHP Phalcon. Using PHP Phalcon, crawl up to 5 pages of our website (). It should output the following: - Number of pages crawled - Number of a unique images - Number of unique internal links - Number of unique external links - Avg page load - Avg word count - Avg Title length - Table to display each page crawled and it's status code

    PHP
    €46 (Avg Bid)
    €46 Průměr. nabídka
    1 nabídky

    We have a system in place to prevent scraping which detects datacenter IPs, then attempts to determine if the IP is a known search crawler or not before blocking the request. Unfortunately, we realized last week it's not actually working as expected because googlebot is being blocked incorrectly. I assume it is something simple and should be a quick resolution for a developer fluent in PHP. $reverseDNS = gethostbyaddr($ip); if( substr($reverseDNS, -strlen('')) != '' || substr($reverseDNS, -strlen('')) != '' || substr($reverseDNS, -strlen('')) != '' || substr($reverseDNS, -strlen('')) != '' || substr($reverseDNS, -strlen('')) != ''

    €28 (Avg Bid)
    €28 Průměr. nabídka
    1 nabídky

    I am Director of a company that provides doctors to hospitals. Safety is important so each week, we need to check and make sure each doctor still has valid registration and there are no problems. I am using Integromat, a JSON based app, to link all my other apps together using a database and invoicing apps. What I would like is an app or module that can connect to my Integromat/Knack database. I want the app to get a NAME or NUMBER from the database and then search a specific webpage for medical registration and automatically populate the search fields. The page has a "print" option. I would like to print the resulting PDF to file and then store that in the module to connect to my database app so it can upload the PDF to the database. I need to check if they are registered (=r...

    €287 (Avg Bid)
    €287 Průměr. nabídka
    9 nabídky

    The project consists of creating a multi-user system that allows users to make specific queries and then "enrich" those queries with additional information. The queries must generate requests for robots that will perform queries on certain sites, the data will be captured and made available to users dynamically. The captured data must be stored in the database. A web interface must be made available for the user to consult, export, delete and etc. A web app admin template will be made available to the frontend. I will send more details to candidates who submit proposals.

    €420 (Avg Bid)
    €420 Průměr. nabídka
    21 nabídky

    Your task is to develop a web page crawler in selenium, which takes the given URL as starting page and crawls all URLs of same second&top-level domain of the website. Means crawl all pages and fetch: - image links - links itself follow recursively all links and buttons fetch of each of the given links and images: - the link-url - all html-attributes so basically you need to - collect all URLs - iterate over all, ensure not to do the crawling-work multiple times - use PageObject pattern and other recommended from seleniumHQ Your input of the function will be a URL and a limiting regex to limit the found URLs for a given regex. The Result of the function will be a jgrapht graph with model classes, which represent the graph-structure of the website. What is NOT n...

    €88 (Avg Bid)
    €88 Průměr. nabídka
    5 nabídky

    You will be building a scraper/crawler that does the following: - Takes input in the format of 'movie title' AND 'year' on request (Eg. Inception 2010) - Goes to and searches the movie - From the search page (eg. ) and matches the exact movie name and year. The second result in this case : - Visit the result page - - Visit the movie player page - - Gets the Google video link (:4f8:c2c:8d65::1&id=bc697d3b2bb1d454&itag=22&source=picasa&begin=0&requiressl=yes&mm=30&mn=sn-4g5edns7&ms=nxu&m

    €13 / hr (Avg Bid)
    €13 / hr Průměr. nabídka
    29 nabídky

    I have tas-crawler for 1 website - it's been working till the site changed few things - need someone to fix it. Do NOT bid if you are NOT proficient with tasks (crawlers). If successful, I'll have more Apify work for you! Details via messaging.

    €22 (Avg Bid)
    €22 Průměr. nabídka
    13 nabídky

    We are a startup Property Management company that manages properties on behalf of the landlords and rent them to visitors by the night on the websites listed on the Title. We want to setup Zoho CRM, but to properly do it we need to use Deluge to create certain functions to extract information from emails such as t...do it we need to use Deluge to create certain functions to extract information from emails such as the text body and the important fields about the apartments reservations. Along with these functions we also need to create an integration which is already half way done using zoho flow to bring information from our main Property Management Software Beds24. The function should use something like regex (or similar) to parse an html. Please contact me for an example fil...

    €181 (Avg Bid)
    €181 Průměr. nabídka
    9 nabídky

    I need a freelancer that can crawl a list of 22,000,000 domains that I will share in four CSV files that you will download from Google Storage. The crawler needs to crawl ONLY THE LANDING PAGE - not the entire site - and capture the following and output a CSV file and stored to Dropbox: 1) Does URL have Google Analytics code - yes or no. Use a search for "Google Analytics" in the source of the page. 2) Is there a link to a privacy policy on the page - yes or no. Use a search for the word "Privacy" in the link text 3) How many unique internal URL links are present on the page. Return href count. 4) Is the URL secure (SSL) - yes or no. 5) Is the URL mobile-friendly - yes or no. Use a search for "meta name="viewport"" in the source of the ...

    €224 (Avg Bid)
    €224 Průměr. nabídka
    11 nabídky

    Your task is to develop a web page crawler in selenium, which takes the given URL as starting page and crawls all URLs of same second&top-level domain of the website. Means crawl all pages and fetch: - image links - links itself follow recursively all links and buttons fetch of each of the given links and images: - the link-url - all html-attributes so basically you need to - collect all URLs - iterate over all, ensure not to do the crawling-work multiple times - use PageObject pattern and other recommended from seleniumHQ Your input of the function will be a URL and a limiting regex to limit the found URLs for a given regex. The Result of the function will be a jgrapht graph with model classes, which represent the graph-structure of the website. What is NOT n...

    €119 (Avg Bid)
    €119 Průměr. nabídka
    8 nabídky

    I need a regular expression to be used in classic ASP to detect garbage text in contact forms. This is a typical example: Ïîëàãàÿñü íà øèðîêèé ïðîèçâîäñòâåííûé. Obviously we cannot filter all extended accented characters as people can contact us from all over the world. The RegEx should only flag excessive consecutive use of accented characters like the example. It should also flag more than 1 hyperlink.

    €106 (Avg Bid)
    €106 Průměr. nabídka
    2 nabídky
    Webcrawler Ukončeno left

    I will provide a freelancer with a specific given website which contains various information on subpages, and will need a automized web crawler which downloads the specified data. It doesn't matter how this data will be stored (whether in a provided database or CSV export), due to this data will only be used for statistical analysis. Summary/Requirement: Specific data from ONE single website shall be provided via download (whether Excel, CSV oder MySQL is irrelevant).

    €113 (Avg Bid)
    €113 Průměr. nabídka
    33 nabídky

    I need a freelancer that can crawl a list of 22,000,000 domains that I will share in four CSV files that you will download from Google Storage. The crawler needs to crawl ONLY THE LANDING PAGE - not the entire site - and capture the following and output a CSV file and stored to Dropbox: 1) Does URL have Google Analytics code - yes or no. Use a search for "Google Analytics" in the source of the page. 2) Is there a link to a privacy policy on the page - yes or no. Use a search for the word "Privacy" in the link text 3) How many unique internal URL links are present on the page. Return href count. 4) Is the URL secure (SSL) - yes or no. 5) Is the URL mobile-friendly - yes or no. Use a search for "meta name="viewport"" in the source of the ...

    €218 (Avg Bid)
    €218 Průměr. nabídka
    7 nabídky
    stocktwits_Irvine Ukončeno left

    Reading in tweet data in SAS. 2 Programs are running, one for each kind of data, but getting too many errors for comfort. Regex code perhaps necessary to read in without errors? Sample data and SAS error log attached.

    €484 (Avg Bid)
    €484 Průměr. nabídka
    2 nabídky