python + 高德地图API实现地图找房

项目简介:根据工作地点信息和58同城爬取的租房信息,通过高德地图进行显示,同时利用高德API自动规划房源到工作地点的通勤路线(公交+地铁)
项目仓库:https://github.com/haohaizhi/58house_spiders

一、数据爬取

# 拉取代码
git clone https://github.com/haohaizhi/58house_spiders.git

完整代码如下:

from bs4 import BeautifulSoup
import requests
import csv
import time
import lxml

def main():

    # 1、筛选条件为租金的范围,可自定义修改
    # 2、地点不同则网址二级域名不同,如此时南京:nj.58.com  北京:bj.58.com
    # 3、因为没有做一些动态切换处理,多次运行可能会触发网址防爬机制,此时需要手动访问该网站进行验证
    # 4、使用此程序需要了解一些爬虫基本原理
    url = "https://nj.58.com/pinpaigongyu/pn/{page}/?minprice=2000_4000"

    # 已完成的页数序号,初时为0

    csv_file = open("house.csv", "w", encoding="utf-8")
    csv_writer = csv.writer(csv_file, delimiter=',')

    for page in range(0,20):
        page += 1
        print("正在爬取第" + str(page) + "页信息........")
        print("URL: ", url.format(page=page))
        time.sleep(3)
        response = requests.get(url.format(page=page))
        response.encoding = 'utf-8'
        html = BeautifulSoup(response.text,features="lxml")
        # lxml 4.0版本
        # html = BeautifulSoup(response.text,features="html.parser")
        house_list = html.select(".list > li")

        for house in house_list:
            house_title = house.select("img")[0]["alt"]
            house_url = house.select("a")[0]["href"]
            house_info_list1 = house_title.split('-')

            # 如果第二列是公寓名或者社区则取第一列作为地址
            if "公寓" in house_info_list1[0] or "社区" in house_info_list1[0]:
                house_info_list2 = house_info_list1[0].split(' ')
                house_info_list2 = house_info_list2[0].replace('【','')
                house_info_list2 = house_info_list2.replace('】', '|')
            else:
                house_info_list2 = house_info_list1[0]
            print(house_info_list2 + " " + house_url)

            house_location = house_info_list2.replace('|',' ')
            house_info_list = house_location.split(' ')
            print(house_info_list)

            house_url = "https://nj.58.com" + house_url
            csv_writer.writerow([house_info_list[1], house_url])
    csv_file.close()

if __name__ == '__main__':
    main()

根据自身需要进行修改如下url:

# 修改代码中url网址 (地区 价格范围)
# 如: 北京:bj.58.com
#     南京:nj.58.com
# 若想进行更准确的筛选,则需要添加其他参数,具体需要参考58同城
url = "https://nj.58.com/pinpaigongyu/pn/{page}/?minprice=2000_4000"

查看所需数据的标签
请添加图片描述
请添加图片描述

#这几行代码就是从网页源码中获取所需数据,这个用到了BeautifulSoup库的select函数,函数具体用法可上网搜索文档,或者在本项目代码仓库的README中查看
		response = requests.get(url.format(page=page))
        response.encoding = 'utf-8'
        html = BeautifulSoup(response.text,features="lxml")
        house_list = html.select(".list > li")

注意:程序运行过程中很大几率会出现中断的情况,基本原因都是爬取网址时出现问题,可能网络波动,可能网站需要人工验证

# 可将部分代码修改成以下方式进行优化
    for page in range(0,20):
     	try:
	        page += 1
	        print("正在爬取第" + str(page) + "页信息........")
	        print("URL: ", url.format(page=page))
	        time.sleep(3)
	        response = requests.get(url.format(page=page))
	        response.encoding = 'utf-8'
	        html = BeautifulSoup(response.text,features="lxml")
	        house_list = html.select(".list > li")
	
	        for house in house_list:
	            house_title = house.select("img")[0]["alt"]
	            house_url = house.select("a")[0]["href"]
	            house_info_list1 = house_title.split('-')
	
	            # 如果第二列是公寓名或者社区则取第一列作为地址
	            if "公寓" in house_info_list1[0] or "社区" in house_info_list1[0]:
	                house_info_list2 = house_info_list1[0].split(' ')
	                house_info_list2 = house_info_list2[0].replace('【','')
	                house_info_list2 = house_info_list2.replace('】', '|')
	            else:
	                house_info_list2 = house_info_list1[0]
	            print(house_info_list2 + " " + house_url)
	
	            house_location = house_info_list2.replace('|',' ')
	            house_info_list = house_location.split(' ')
	            print(house_info_list)
	
	            house_url = "https://nj.58.com" + house_url
	            csv_writer.writerow([house_info_list[1], house_url])
        except:
           	pass
    csv_file.close()

请添加图片描述
请添加图片描述
刚开始进行调试时可以先尝试只分析某一页的html源码

# 代码如下
from bs4 import BeautifulSoup
import requests
import csv
import time
import lxml

def main():
    url = "https://nj.58.com/pinpaigongyu/pn/1/?minprice=2000_4000"
    response = requests.get(url.format(page=page))
    response.encoding = 'utf-8'
    print(response)
    
if __name__ == '__main__':
    main()

将执行后打印的内容Ctrl + A, Ctrl + C 复制粘贴到临时文件中,如test.txt
然后就可以针对这个临时文件进行后续数据提取的操作,等程序调试的没问题了,在用最最终版的代码

from bs4 import BeautifulSoup
import requests
import csv
import time
import lxml
def main():
	csv_file = open("house.csv", "w", encoding="utf-8")
    csv_writer = csv.writer(csv_file, delimiter=',')
    
	f = open('./test.txt', 'r')
    response = f.read()
    f.close()
	
	html = BeautifulSoup(response,features="lxml")
    house_list = html.select(".list > li")
    for house in house_list:
        house_title = house.select("img")[0]["alt"]
        house_url = house.select("a")[0]["href"]
        house_info_list1 = house_title.split('-')

        # 如果第二列是公寓名或者社区则取第一列作为地址
        if "公寓" in house_info_list1[0] or "社区" in house_info_list1[0]:
            house_info_list2 = house_info_list1[0].split(' ')
            house_info_list2 = house_info_list2[0].replace('【','')
            house_info_list2 = house_info_list2.replace('】', '|')
        else:
            house_info_list2 = house_info_list1[0]
        print(house_info_list2 + " " + house_url)

        house_location = house_info_list2.replace('|',' ')
        house_info_list = house_location.split(' ')
        print(house_info_list)

        house_url = "https://nj.58.com" + house_url
        csv_writer.writerow([house_info_list[1], house_url])
    csv_file.close()
    
if __name__ == '__main__':
    main()    

二、高德地图显示

打开该html文件
在这里插入图片描述
初始时坐标中心为北京,输入工作地点并回车后会进行跳转
在这里插入图片描述
将house.csv文件导入,点击蓝色图标,将自动规划通勤路线
请添加图片描述
点击房源链接地址后将跳转到指定网址
请添加图片描述
到此,整个项目已介绍完毕。

已同步发表到个人网站:https://blog.mehoon.com/269.html

Logo

为开发者提供学习成长、分享交流、生态实践、资源工具等服务,帮助开发者快速成长。

更多推荐