追加中原地产相关信息

  • 全市均价
  • 全市均价环比上月
  • 全市均价同比去年
  • 全市在售房源
  • 全市近60天成交量
  • 以上信息浦东,嘉定,奉贤

In [2]:
import requests,time,os,csv,re,logging
from bs4 import BeautifulSoup
import numpy as np
import traceback

from lianjia_requests import get_n_times

In [37]:
ex_date = time.strftime('%Y%m%d')
ex_time = time.strftime('%H:%M:%S')

In [ ]:
print(ex_date + ex_time+ ":正在获取...")

In [3]:
urls_dict={"lianjia":"http://sh.lianjia.com/ershoufang",
           "Shanghai":"http://sh.centanet.com/chengjiao/contrast/?cestcode=&regionid=0&gscopeid=0",
           "pudong":"http://sh.centanet.com/chengjiao/contrast/?cestcode=&regionid=2172&gscopeid=0",
           "pudong_chuansha":"http://sh.centanet.com/chengjiao/contrast/?cestcode=&regionid=2172&gscopeid=217218",
           "jiading":"http://sh.centanet.com/chengjiao/contrast/?cestcode=&regionid=2178&gscopeid=0",
           "fengxian":"http://sh.centanet.com/chengjiao/contrast/?cestcode=&regionid=2180&gscopeid=0"}

In [ ]:
fieldnames=['ex_date','ex_time', 
            'lj_on_sale','lj_90day_deal', 'lj_yesterday_viewed',
           'sh_on_sale','sh_60day_deal','sh_ave_p','sh_huanbi','sh_tongbi',
           'pd_on_sale','sh_60day_deal','sh_ave_p','sh_huanbi','sh_tongbi']

In [ ]:


In [ ]:


In [ ]:


In [ ]:


In [ ]:


In [ ]:


In [38]:
url = 'http://sh.lianjia.com/ershoufang'
html =get_n_times(url, n=6)

In [42]:
soup=BeautifulSoup(html,'html.parser')
spans=soup.find_all('span',attrs={'class':'num strong-num'})

total_on_sale=spans[0].string
last_90day_deal = spans[1].string
yesterday_viewed = spans[2].string

In [45]:
item={'ex_date':ex_date,
      'ex_time':ex_time, 
      'total_on_sale':total_on_sale,
      'last_90day_deal':last_90day_deal,
      'yesterday_viewed':yesterday_viewed}

In [51]:
path=os.path.join(os.getcwd(),'results/lianjia_trends.csv')

fieldnames=['ex_date','ex_time', 'total_on_sale','last_90day_deal', 'yesterday_viewed']

if  not os.path.exists(path):     #如果文件不存在,创建文件并写入表头
        csv_file=open(path,'w',newline='')
        writer=csv.DictWriter(csv_file,fieldnames=fieldnames)
        writer.writeheader()
        csv_file.close()
        
csv_file=open(path,'a+',newline='')
writer=csv.DictWriter(csv_file,fieldnames=fieldnames)
writer.writerow(item)
csv_file.close()

In [56]:
n=3
print('获取结束,%d秒后关闭' %n)  
time.sleep(n)


获取结束,3秒后关闭

In [ ]: