java实现通过url获取状态码与页面内容

package httptest.demo.test1;

import java.io.BufferedReader;
import java.io.DataOutputStream;
import java.io.IOException;
import java.io.InputStreamReader;
import java.io.UnsupportedEncodingException;
import java.net.HttpURLConnection;
import java.net.MalformedURLException;
import java.net.URL;

public class demo1 {
	//添加url
public static final String Add_url="http://xiaojingjing.iteye.com/";

public static void addpost() throws IOException{
	//创建连接
	try {
		URL url=new URL(Add_url);
		HttpURLConnection connection =(HttpURLConnection)url.openConnection();
		connection.setDoOutput(true);
		connection.setDoInput(true);
		connection.setRequestMethod("POST");
		connection.setUseCaches(false);
		connection.setInstanceFollowRedirects(true);
		connection.setRequestProperty("Content-Type","application/x-www-form-urlencoded");
		connection.connect();
		
		//POST请求
		DataOutputStream out=new DataOutputStream(connection.getOutputStream());
//		out.writeBytes("email");
		out.flush();
		out.close();
		//读取响应
		BufferedReader reader=new BufferedReader(new InputStreamReader(connection.getInputStream()));
		String lines;
		StringBuffer sb=new StringBuffer("");
		while((lines = reader.readLine())!=null){
			lines =new String(lines.getBytes(),"utf-8");
			sb.append(lines);
		}
		System.out.println(sb);
		reader.close();
		connection.disconnect();
	 } catch (MalformedURLException e) {
         // TODO Auto-generated catch block
         e.printStackTrace();
     } catch (UnsupportedEncodingException e) {
         // TODO Auto-generated catch block
         e.printStackTrace();
     } catch (IOException e) {
         // TODO Auto-generated catch block
         e.printStackTrace();
     }
}
public static void main(String args[]) throws IOException
{
	addpost();
	}
}

 如果网站屏蔽了“爬虫”,就会获取不到网页内容,只能返回一个状态码。

猜你喜欢

转载自xiaojingjing.iteye.com/blog/2261655