1. to use urllib2/sgmllib,list all URLs on a web page: import urllib2 from sgmllib import SGMLParser class URLLister(SGMLParser): def reset(self): ...
by xxdxxd - Python文档中心 - 2008-11-14 09:31:53 阅读(2216) 回复(0)
import re,urllib2 from urllib import urlopen def get(): doc1=urllib2.urlopen("http://money.finance.sina.com.cn/corp/go.php/vFD_BalanceSheet/stockid/002024/ctrl/2013/displaytype/4.phtml") 如果我要对其中的“002024”进行用户输入替换的话应该怎么办
如题,想实现模拟用户登录,可是看着结果好像是跳转失败! 老大们给看看是啥原因 下边是我用httpfox抓的包 系统:Red Hat Enterprise Linux Server release 6.0 python版本: python 2.6.5 #!/usr/bin/python import HTMLParser import urlparse import urllib import urllib2 import cookielib import string import re posturl = 'http://10.17.16.66/designs/imm/index.php' cj = cookielib.LWPCookieJar() cookie_s...
[code]import urllib2 # Create an OpenerDirector with support for Basic HTTP Authentication... auth_handler = urllib2.HTTPBasicAuthHandler() auth_handler.add_password(realm='PDQ Application', uri='https://mahler:8092/site-updates.py', user='klem', passwd='kadidd!ehopper') opener = urllib2.build_opener(auth_handler) # ...a...
最近正在学习python的urllib2模块,碰到了个问题 例如cnbird.php 然后我们用python来设置 >>> import urllib >>> import urllib2 >>> url = 'http://127.0.0.1/cnbird.php' >>> values = {'cnbird' : ''} >>> data = urllib.urlencode(values) >>> req = urllib2.Request(url, data) >>> response = urllib2.urlopen(req) >>> the_page = response.read() >>> print ...
http://blog.alexa-pro.cn/?p=195 简介: urllib2是python的一个获取url(Uniform Resource Locators,统一资源定址器)的模块。它用urlopen函数的形式提供了一个非常简洁的接口。这使得用各种各样的协议获取url成为可能。它同时 也提供了一个稍微复杂的接口来处理常见的状况-如基本的认证,cookies,代理,等等。这些都是由叫做opener和handler的对象来处理的。 以下是获取url最简单的方式: import urllib2 response = url...
#!/usr/bin/env python import urllib2 # change followings before use user = 'foo' passwd = 'bar' proxyserver = '1.2.3.4:5' url = 'http://www.google.com/' def proxy1(): # work proxy = 'http://%s:%s@%s' % (user, passwd, proxyserver) opener = urllib2.build_opener( urllib2.ProxyHandler({'http':proxy}) ) urllib2.install_opener( opener ) sContent = urllib2.urlopen(url) print ...
查看文章 python urllib2技术补充参考--中文版01 2008-12-12 17:38 python urllib2技术补充参考--中文版01 此文由我( 54snapple )翻译自 http://www.voidspace.org.uk/python/articles/urllib2.shtml 。原文标题:urllib2 - The Missing Manual -副标题:HOWTO Fetch Internet Resources with python。英文版权归文章原作者所有。转载请保留以上信息。 行文仓促,未作校正,请谅解指正。 -----------------------------------...
#!/usr/bin/python # -*- coding:utf-8 -*- import urllib2 header = {'User-Agent':'Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.1.6) Gecko/20091201 Firefox/3.5.6'} req = urllib2.Request( url= "https://www.google.com.hk/search?num=100&hl=zh-CN&newwindow=1&safe=strict&q=inurl%3Aadmin_login.aspx&oq=inurl%3Aadmin_login.aspx&gs_l=serp.3...125521.131943.0.132041.38.31.1.0.0.3.209.2367.23j3...
如题,文件可以用多线程读出来 ,但是读出后 还是1个个的去检查,还请帮忙 web.py #!/usr/bin/env python #coding=utf-8 import urllib import urllib2 from time import ctime from checkfd import check_report def check_url(url_name): try: h=urllib2.HTTPHandler(debuglevel=0) opener = urllib2.build_opener(h) request = urllib2.Request(url_name) feeddata = opener.open(reque...