Pengembangan Aplikasi Java Web Crawler

Elfianto, Dwi Riyanti (0472212) (2008) Pengembangan Aplikasi Java Web Crawler. Undergraduate thesis, Universitas Kristen Maranatha.

[img]
Preview
Text
0472212_Abstract_TOC.pdf - Accepted Version

Download (85Kb) | Preview
[img]
Preview
Text
0472212_Chapter1.pdf - Accepted Version

Download (113Kb) | Preview
[img] Text
0472212_Chapter2.pdf - Accepted Version
Restricted to Registered users only

Download (124Kb)
[img] Text
0472212_Chapter3.pdf - Accepted Version
Restricted to Registered users only

Download (211Kb)
[img] Text
0472212_Chapter4.pdf - Accepted Version
Restricted to Registered users only

Download (1823Kb)
[img] Text
0472212_Chapter5.pdf - Accepted Version
Restricted to Registered users only

Download (651Kb)
[img]
Preview
Text
0472212_Conclusion.pdf - Accepted Version

Download (77Kb) | Preview
[img]
Preview
Text
0472212_Cover.pdf - Accepted Version

Download (499Kb) | Preview
[img]
Preview
Text
0472212_References.pdf - Accepted Version

Download (55Kb) | Preview

Abstract

This project primarily exists as my personal desire to create a web crawler application for browsing the web and store the information without much interaction from human. And Java is choses as development base because it runs very well on many platform (including my Linux rig) and it has a good collection of network or Internet ready API which fulfill my needs. The great threading ability of Java also becoming a major advantage for this project because web crawler ability to process multiple link simultaneously must be implemented using threads. Web crawler means any application which capable of automatically process any hyper links found on a HTML page and process it to gather more links which will be processed again, inside the application this behavior will take place until a stop condition is met (usually a certain link depth or number of files retrieved). Although there are many purpose of a web crawler, the main purpose of using it is for information gathering or archiving. The contents within this paper will describe the process of developing a web crawler used for archiving files gathered from HTML paged to local storage for off line browsing. And for the final words, this project has meets it's goal though not 100% perfect because there's still some minor glitch in few of the code. Although rare, this problem will cause inability to retrieve some type of relative links. In the end I just going to say thanks to all who has support this project, I'm looking forward to fix the glitches and make the application better.

Item Type: Thesis (Undergraduate)
Uncontrolled Keywords: Web Crawler, Java Network Programming, Spider
Subjects: T Technology > T Technology (General)
Divisions: Faculty of Information Technology > 72 Information Technology Department
Depositing User: Perpustakaan Maranatha
Date Deposited: 28 Sep 2015 10:56
Last Modified: 28 Sep 2015 10:56
URI: http://repository.maranatha.edu/id/eprint/15612

Actions (login required)

View Item View Item