Network Programming, Simple Browser

Network Programming, Simple Browser

With Java, you can easily write a client program that connects to a Web server and downloads an HTML file. You can also easily expand this simple program into a Web crawler by parsing the downloaded file, extracting the links referenced in that file, and using similar code to download the referenced files from the servers identified by those links. But be careful–you might download the whole Web onto your computer!

This program displays a specified HTML file on a specified server in raw text format:

 import*;import*;class SocketHttp01{	public static void main(String[] args){		String server = "";//server name		int port = 80; //http port		try{			Socket socket = new Socket(server,port);//get socket			//Get input and output streams from the socket      			BufferedReader inputStream = 					new BufferedReader(new InputStreamReader(							socket.getInputStream()));			PrintWriter outputStream = 					new PrintWriter(new OutputStreamWriter(							socket.getOutputStream()),true);			outputStream.println("GET /scoop/2train.htm");			//GET a specified file			String line = null;			while((line = inputStream.readLine()) != null) 				System.out.println(line);			//Loop reading and displaying lines.		socket.close();		}//end try		catch(UnknownHostException e){			System.out.println(e);			System.out.println(					"Connection Failure. Are you online?");		}//end catch UnknownHostException		catch(IOException e){System.out.println(e);}	}//end main}//end class SocketHttp01
See also  5 Tips for Choosing the Right Medical Software for Your Practice

About Our Editorial Process

At DevX, we’re dedicated to tech entrepreneurship. Our team closely follows industry shifts, new products, AI breakthroughs, technology trends, and funding announcements. Articles undergo thorough editing to ensure accuracy and clarity, reflecting DevX’s style and supporting entrepreneurs in the tech sphere.

See our full editorial policy.

About Our Journalist