Web Crawler Jobs

75 were found based on your criteria {{ paging.total|number:0 }} were found based on your criteria

show all
  • Hourly ({{ jobTypeController.getFacetCount("0")|number:0}})
  • Fixed Price ({{ jobTypeController.getFacetCount("1")|number:0}})
Hourly - Expert ($$$) - Est. Time: Less than 1 week, Less than 10 hrs/week - Posted
Move my WordPress site: lpnbsn . net to another domain name. Must have experience moving sites. - Change all internal links? - 301 the site to the new site? - What to do with current redirects? - How to inform Google about change? - Additionally, how to redirect the best of the incoming links, if that is possible? Regards, Jeff
Skills: Web Crawler HTML Web Hosting Web Testing
Fixed-Price - Intermediate ($$) - Est. Budget: $100 - Posted
****Desired Skills Lead Generation, Market Research, Data Mining, Data scraping**** We are looking to create a list of potential clients using different sources such as white pages, google search etc. to complete list of Shippers, Distributors, Manufactures and exporters (Pharmaceuticals, electronics, perfume design ,cosmetics for furniture ,airspace ATVs motorcycles , quadricycle ,energy supplies, grinders drills drilling manufacturers oil/gas energy , military) Business Name , Email, Contact Name (first and last name) , Phone Number Business, Address(street,city,state zip) ,company website URL Additional info: years in business number of employees facebook page specialization, Carrier Route Put aggregate data into CSV or excel format
Skills: Web Crawler Data mining Data scraping Google search
Fixed-Price - Expert ($$$) - Est. Budget: $500 - Posted
For our Irish based digital music distribution software company (www.EVEARA.com) we are looking for a extremely innovative, very creative, very organized, and skillful website developer. We planning to launch our new website in March 2016, and are looking for something really outstanding.
Skills: Web Crawler Adobe Photoshop AJAX CSS
Hourly - Expert ($$$) - Est. Time: More than 6 months, Less than 10 hrs/week - Posted
If you understand what UIMA is and you have built an automated scalable web crawler before with custom annotators and dictionaries, we want to hear from you. Seeking experienced UIMA developer to help scope out the architecture for and then lead the effort to build a domain specific UIMA-based scalable crawler that can be used to extract data from tens of thousands of websites in a session, powered by multiple servers. Projects are confidential, to be discussed under NDA with qualified candidates.
Skills: Web Crawler Web Crawling Data mining Data Modeling
Fixed-Price - Entry Level ($) - Est. Budget: $500 - Posted
Proposals requested for data driven ‘search’ Section 1: We are looking to create a ‘search engine’ for our design process that generates results based on the user’s queries with them having entered specific keywords. This would include the areas of artificial intelligence, machine intelligence, data mining, data scraping and not just simply web development and scraping. We would want the results to recommending specific institutions based on the user query. These results would be displayed in a fixed generic format, but content would be based on user’s unique query. Every subsequent query by the user which is dependent on an initial query will adopt the same search format until the results generated for the user are enough to meet the user’s needs. The functionalities are driven by machine learning algorithms, users’ keywords and information on institutions in the news which will serve as pointers for the algorithm to decide on the best matching institutions. Decision making will be dynamic and tailored uniquely to the individual. Section 2: This should then allow the user to select a number of the generated results, namely institutions that meet the needs of the query, the algorithm should then be able to scrape information on the specific institutions and provide the information to users, this includes management team, contact details, anything available. This helps to narrow down users’ options and provides them with a strong basis to work from. Service Provider Skillset: Data Scientist/Programmer, Web/App Developer and Web Tool Specialist and in order to better assess your suitability for the project, we would like to see a portfolio of projects you have completed related to data mining, artificial intelligence and machine learning. Interested Service Providers should send a brief EOI stating the following: 1. Their understanding of the project and demonstration of experience, expertise and capacity to do the job with references to similar jobs completed if any 2. Proposed time frame and cost (fixed price fee) to accomplish the project
Skills: Web Crawler Algorithms Data mining Data Science
Hourly - Intermediate ($$) - Est. Time: 1 to 3 months, 10-30 hrs/week - Posted
This job is focused on advancement of the experience that thousands of users get navigating, browsing, searching and comparing the content offered through our proprietary technology platform. The end-result (output of the ontology model) will be a set of intuitive and comprehensive multi-level navigation structures (hierarchical taxonomies, facets) for browsing, searching and tagging the content offered to our clients. The end-task is envisioned to be primarily achieved with the usage of Semantic Web concepts and data (LOD and other available SKOS) as per Semantic Web standards. The task most likely will require knowledge/learning of several RDF-based schemas (Resume RDF, HRM Ontology, HR-XML, FOAF, SCIOC, Schema.org) and usage of the W3C’s Semantic Web technology stack components (SPARQL, Protege, Semantic resoners). Key tasks: - Definition of RDF Schema and ontologies based on several existing RDF Schemas (Resume RDF, HRM Ontology, HR-XML, FOAF, SCIOC, Schema.org, etc.) - linking available LOD and SKOS data sets, building several core multi-level hierarchical taxonomies (magnitude of tens of thousands of elements) comprehensively describing the content in our system - Rule-based processing and linking of multiple existing, as well as obtained sets of data using semantic reasoners - Definition, structuring and optimization of hierarchical data sets, definition and maintenance of hierarchical relationships of particular terms (facets) - Research (independent, as well as guided by management team) on publicly available SKOS and LOD sets related to the content of the platform from public (international standards, patent databases, public and government databases, various organizational, available XML datasets, etc.), as well as acquired proprietary sources - Retrieval and ETL of multiple additional data sets from multiple sources - Tagging, Classification, entity extraction - Working with management team to maintain and advance particular segments of defined taxonomies Optional Stretch-Tasks (Depending on Candidate's Qualifications): - Automatic analysis of content, extraction of semantic relationships - Auto-tagging, auto-indexing - Integration and usage of selected IBM Watson services for content analysis - Integration with Enterprise Taxonomy Management platforms (Mondeca, Smartlogic, PoolParty, or others) This job will initially require commitment of 15-20 hours per week over 3-6 months engagement. Interaction with a responsible manager will be required at least twice a week over Skype and Google Hangouts. Longer-term cooperation is possible based on the results of the initial engagement. Required Experience: - Detailed knowledge of Semantic Web concepts and techniques - Intimate familiarity with W3C’s Semantic Web technology stack (RDF, SPARQL, etc.) - Hands-on experience with LOD (DB Pedia and others) and various SKOS - Experience of modeling data based on various RDF schemas (Resume RDF, HRM Ontology, HR-XML, FOAF, SCIOC, ISO 25964, etc.) - Knowledge of common open-source ontology environments and tools (Mediawiki, Protege, etc.) or other enterprise-grade ontology tools (Synaptica, DataHarmony, PoolParty, Mondeca, Top Braid, etc.) - Experience of work with semantic reasoners - Prior experience of content management and maintenance of taxonomies for consumer or e-commerce applications Additional Preferred Experience: - Background in Library and Information Science (MLIS), Knowledge Management, Information Management, Linguistics or Cognitive Sciences - Familiarity with common classification systems - Experience working with catalog and classification systems and creation of thesauri - Auto-tagging, auto-classification, entity extraction
Skills: Web Crawler Web Crawling Data Analytics Data Entry