this is the error in the following python code, can any one help me error{Traceback (most recent call last): File "C:\Python27\Scripts\BeOk\getBeOKExperts.py", line 6, in <module> from BeautifulSoup import BeautifulSoup ImportError: No module named BeautifulSoup}
"#encoding=utf8 from codecs import open from collections import defaultdict import re from BeautifulSoup import BeautifulSoup import mechanize import cookielib import html2text import time def getbr(): br = mechanize.Browser() # Cookie Jar cj = cookielib.LWPCookieJar() br.set_cookiejar(cj) # Browser options br.set_handle_equiv(True) br.set_handle_gzip(True) br.set_handle_redirect(True) br.set_handle_referer(True) br.set_handle_robots(False) # Follows refresh 0 but not hangs on refresh > 0 br.set_handle_refresh(mechanize._http.HTTPRefreshProcessor(), max_time=1) # User-Agent (this is cheating, ok?) br.addheaders = [('User-agent', 'Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.0.1) Gecko/2008071615 Fedora/3.0.1-1.fc9 Firefox/3.0.1')] return br def logthis(text): open("log.txt","a","utf8").write(text+"\n") def getCommunity(community,url,out=""): # Browser # The site we will navigate into, handling it's session i = 1 flag = True discussions = [] baseDiscussion = [] while flag: print i currurl = url+"/"+str(i) try: br = getbr() br.open(currurl) #br.follow_link(text='link') html = br.response().read() soup = BeautifulSoup(html) if soup.find("title").string == u'\r\n\t\u05d4\u05d5\u05d3\u05e2\u05ea \u05de\u05e2\u05e8\u05db\u05ea - BeOK\r\n': print "done at ",i,community logthis("done at "+str(i)+" "+community) return True hrefList = soup.findAll('div',{"class":"MsgTtlChildRow"}) print currurl #print hrefList for link in hrefList: #print str(link) #continue span = link.find('div',{"class":"MsgUsr"}) if "frm_mngr" in str(span): mgr = span.find("span",{"class":"frm_mngr"}).string if not "''" in mgr: continue mgr = mgr.replace("'","") date = link.find('span',{"class":"MsgDate"}).string.split(" ")[1] #out.write(community+"\t"+mgr+"\t"+date+"\n") print community.rstrip(),date,mgr #fout = open("corpus\\"+community+"-"+date+"-"+mgr,"w","utf8") ansDiv = link.nextSibling.find('div',{"class":"BodyMesInner"}) print "bla" ans = fixHtml2(str(ansDiv)) print "bla" print ans #fout.write(fixHtml(link.find('div',{"class":"BodyMesInner"}).string)+"\n") #fout.close() questionDiv = link.previousSibling.find('div',{"class":"BodyMesInner"}) print "bla",questionDiv quesiton = fixHtml2(str(questionDiv)) print question span = None soup = None br = None except: time.sleep(60) i+=1 return list(set(discussions)) def fixHtml(page): page = page.replace("</p>","\n") page = page.replace("</P>","\n") page = page.replace("<br />","\n") page = page.replace("<BR />","\n") page = page.replace("<br>","\n") page = page.replace("<BR>","\n") page = page.replace(""","'") reg = re.compile("<") reg2 = re.compile(">") page = " ".join([x[-1] for x in map(reg2.split,reg.split(page))]) page = page.replace("\r\n\t\t\t","\n") return page def fixHtml2(page): page = page.split('ner">')[1].split("<div")[0] print page page = page.replace("</p>","\n") page = page.replace("</P>","\n") page = page.replace("<br />","\n") page = page.replace("<BR />","\n") page = page.replace("<br>","\n") page = page.replace("<BR>","\n") page = page.replace(""","'") return page def getText(br,url): br.open(url) html = br.response().read() soup = BeautifulSoup(html) title = fixHtml(soup.find('h1',{'class':"articleName"}).contents[0]) #print title artics = soup.findAll('div',{'class':"article"}) text = "\n"+fixHtml(str(artics[0]).split('"article">')[1].split('</div>')[0]) text += "\n<EXPERT>"+ fixHtml(str(artics[1]).split('"article">')[1].split('</div>')[0])+"</EXPERT>" text = text.decode("utf-8") #text = artics[0] + #print type(title),type(text) return title+text def getForums(file = "links.htm"): #out = open("beokDates","w","utf8") soup = BeautifulSoup(open(file,"r").read()) communities = soup.findAll("a",{"class":"MainList"}) for comm in communities: #print comm["href"] getCommunity(comm.string,comm["href"]) getForums() #links = getQALinks() file = "links.htm" soup = BeautifulSoup(open(file,"r").read()) comm = soup.findAll("a",{"class":"MainList"})[0] br = getbr() currurl = comm["href"]+"/3" br.open(currurl) html = br.response().read() soup = BeautifulSoup(html) hrefList = soup.findAll('div',{"class":"MsgTtlChildRow"})[0] " -- https://mail.python.org/mailman/listinfo/python-list