如何以Java编程方式下载网页 [英] How do you Programmatically Download a Webpage in Java
本文介绍了如何以Java编程方式下载网页的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!
问题描述
我想要能够获取网页的html并将其保存到 String
,所以我可以做一些处理。此外,我如何处理各种类型的压缩。
I would like to be able to fetch a web page's html and save it to a String
, so I can do some processing on it. Also, how could I handle various types of compression.
我将如何使用Java?
How would I go about doing that using Java?
推荐答案
这里有一些测试的代码使用Java的 URL 类。我建议做一个比我在这里处理异常或将它们传递给调用堆栈更好的工作。
Here's some tested code using Java's URL class. I'd recommend do a better job than I do here of handling the exceptions or passing them up the call stack, though.
public static void main(String[] args) {
URL url;
InputStream is = null;
BufferedReader br;
String line;
try {
url = new URL("http://stackoverflow.com/");
is = url.openStream(); // throws an IOException
br = new BufferedReader(new InputStreamReader(is));
while ((line = br.readLine()) != null) {
System.out.println(line);
}
} catch (MalformedURLException mue) {
mue.printStackTrace();
} catch (IOException ioe) {
ioe.printStackTrace();
} finally {
try {
if (is != null) is.close();
} catch (IOException ioe) {
// nothing to see here
}
}
}
这篇关于如何以Java编程方式下载网页的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!
查看全文