如何获取当前从主机访问的URl列表,包括HTTP,HTTPS以及其他协议 [英] how to get the list of URl currently accessing from a host both the HTTP , HTTPS, other protocols also

查看:253
本文介绍了如何获取当前从主机访问的URl列表,包括HTTP,HTTPS以及其他协议的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!

问题描述

我很麻烦.我需要代码或一些想法来获取使用浏览器(如mozzila或IE或Chrome等)访问url的列表.过滤方法.

请帮帮我.

在此先谢谢您.

I am in big trouble.I need the code or some idea to get the list of url''s accessing using the Browser like mozzila or IE or Chrome ,etc..In the real time so that i can look into some filtering methods.

Please help me.

Thanks in Advance.

推荐答案

不可能获得HTTP或HTTPS列表(以及类似的FTP,SFTP,FTPS等)(除非有)您只需尝试所有可能的字符组合,这几乎是不实际的:-));并且有关其他协议"的问题不正确.

这很容易理解.假设您已经有一个工作站点,只需在HTTP服务器提供的服务器计算机上的目录下创建一个新的HTML文件.这使得新的URL可以立即通过HTTP或HTTPS访问,而无需重新启动服务器.添加此文件不会触发任何操作. HTTP服务器就是这样工作的.您没有添加对此文件的引用.没有包含此资源的列表或其他容器.关于此文件存在的唯一信息是文件本身.在正常且安全配置的服务器中,也无法从客户端访问目录列表.

(纯粹是推测,您可以想象某些不寻常的服务会提供所有可用URL的索引;可以实现;但这不是通常的做法.因为没有足够的服务,所以不使用此类服务​​.实际的理由.这就是Web爬网程序做如此艰苦的工作的原因-它们仅通过可用Web链接的图形进行传播.)

关于FTP的所有变体的相同故事.没办法.

—SA
It is impossible to get a list of HTTP or HTTPS (and the similar: FTP, SFTP, FTPS, etc.) accessible URLs (unless you simply try all possible combinations of characters, which is hardly practical :-)); and the question about "other protocol" is incorrect.

This is easy to understand. Imagine you already have a working site and simply create a new HTML file under the directory on the server computer served by an HTTP server. This makes the new URL immediately accessible via HTTP or HTTPS without re-starting the server. Nothing is triggered by adding this file. This is how the HTTP server works. You did not add a reference to this file; there is no a list or other container which contain this resource. The only information about presence of this file is the file itself. In a normally and safely configured server the directory listing is also not accessible from the client.

(As a pure speculation, you can imagine some unusual service which provides the index of all available URLs; it would be possible to implement; but this this is not what is done in common practice. Such services are not used because there is no enough practical justification. That''s why the Web crawlers do such a hard job — they only travel by the graph of available Web links.)

Same story about all variants of FTP. There is no way.

—SA


这篇关于如何获取当前从主机访问的URl列表,包括HTTP,HTTPS以及其他协议的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!

查看全文
相关文章
登录 关闭
扫码关注1秒登录
发送“验证码”获取 | 15天全站免登陆