from selenium import webdriver
from selenium.webdriver.common.keys import Keys
from bs4 import BeautifulSoup
import urllib3
import urllib.request
import requests
import re
import queue
from bs4 import BeautifulSoup
import scrapy
'''
browser = webdriver.PhantomJS()
browser.get('http://www.aae.wisc.edu/')
print(browser.title)
'''
faculty_name = []()
faculty={}
driver = webdriver.PhantomJS()
driver.get("http://www.cs.wisc.edu/people/faculty/")
soup = BeautifulSoup(driver.page_source,"html.parser")
soup.prettify()
a=0
for item in soup.find_all(class_="views-field views-field-field-full-name"):
print(item.contents[1]().string)
a=a+1
print(a)
print("************************")
b=0
for item in soup.find_all(class_="views-field views-field-field-title"):
print(item.contents[1]().string)
b=b+1
print(b)
c=0
personal_url=[]()
for item in soup.find_all(class_="views-field views-field-field-full-name"):
print("www.cs.wisc.edu"+item.a['href']())
personal_url.append("http://www.cs.wisc.edu"+item.a['href']())
c=c+1
print(c)
i=0
while i<1:
print(personal_url[i]())
response = urllib.request.urlopen(personal_url[i]())
content = response.read()
soup = BeautifulSoup(content, "html.parser")
for item in soup.find_all(re.compile("^Personal Website:")):
print(item.a)
i=i+1
近期评论