SM01: Web Crawling with JLC parallel execution pattern - experiment reports

Published: 1 Nov 2017 | Version 1 | DOI: 10.17632/5prx8vycr5.1
Contributor(s):

Description of this data

Research project SM01 (Parallel Semantic Crawler for manufacturing multilingual web...)

The JLC pattern is running parallel thread for each domain in the Crawl Job domain cue.

Objective of the experiment was to find what number of parallel DLC threads (TCmax) results in the highest execution efficiency. The crawlers were run: 2, 4, 6, 8, 10, 12, 14 and 16 parallel DLC threads, against the Sc subset. Other configuration parameters: Load Take per iteration (LT) set to 1 and Page Loads limit (PLmax) set to 30. Sampling period of the resource utilization: 5 seconds.

Experiment data files

Related links

Latest version

  • Version 1

    2017-11-01

    Published: 2017-11-01

    DOI: 10.17632/5prx8vycr5.1

    Cite this dataset

    Grubić, Goran (2017), “SM01: Web Crawling with JLC parallel execution pattern - experiment reports”, Mendeley Data, v1 http://dx.doi.org/10.17632/5prx8vycr5.1

Institutions

University of Belgrade Faculty of Organizational Sciences

Categories

Manufacturing Industry, Serbian Language, Web Site

Mendeley Library

Organise your research assets using Mendeley Library. Add to Mendeley Library

Licence

CC BY 4.0 Learn more

The files associated with this dataset are licensed under a Creative Commons Attribution 4.0 International licence.

What does this mean?

This dataset is licensed under a Creative Commons Attribution 4.0 International licence. What does this mean? You can share, copy and modify this dataset so long as you give appropriate credit, provide a link to the CC BY license, and indicate if changes were made, but you may not do so in a way that suggests the rights holder has endorsed you or your use of the dataset. Note that further permission may be required for any content within the dataset that is identified as belonging to a third party.

Report