Google Maps crawler using Selenium

Overview

Google Maps Crawler using Selenium

antifragile project python version GitHub Code style: black try/except style: tryceratops Open in Visual Studio Code Follow guilatrova

Built as part of the Antifragile Dev Project

Selenium crawler that browses Google Maps as a regular user and stores the data in an object.


Sample

Sample

Extracted data example:

 Place(                                                                                                            │ │
│ │ │   name='Pizza Me Santos',                                                                                    │ │
│ │ │   address='Av. Washington Luis, 565 - loja 05 - Boqueirão, Santos - SP, 11055-001',                          │ │
│ │ │   business_hours={                                                                                           │ │
│ │ │   │   'Wednesday': '6–10:30PM',                                                                              │ │
│ │ │   │   'Thursday': '6–10:30PM',                                                                               │ │
│ │ │   │   'Friday': '6–11PM',                                                                                    │ │
│ │ │   │   'Saturday': '6–11PM',                                                                                  │ │
│ │ │   │   'Sunday': '6–10:30PM',                                                                                 │ │
│ │ │   │   'Monday': '6–10:30PM',                                                                                 │ │
│ │ │   │   'Tuesday': '6–10:30PM'                                                                                 │ │
│ │ │   },                                                                                                         │ │
│ │ │   photo_link='https://lh5.googleusercontent.com/p/AF1QipMyVkKioODaU0A_ogHPXosm_QcMndZN6I6YHIDo=w408-h272-k-no│ │
│ │ │   rate='5.0',                                                                                                │ │
│ │ │   reviews='16 reviews',                                                                                      │ │
│ │ │   extra_attrs={                                                                                              │ │
│ │ │   │   'Menu': 'Menu\npizzame-santos.goomer.app',                                                             │ │
│ │ │   │   'Website: pizzame-santos.goomer.app ': 'pizzame-santos.goomer.app',                                    │ │
│ │ │   │   'Phone: (13) 3385-0059 ': '(13) 3385-0059',                                                            │ │
│ │ │   │   'Plus code: 2MHC+WF Boqueirão, Santos - State of São Paulo': '2MHC+WF Boqueirão, Santos - State of São │ │
│ │ Paulo'                                                                                                         │ │
│ │ │   },                                                                                                         │ │
│ │ │   traits={                                                                                                   │ │
│ │ │   │   'Service options': ['No-contact delivery', 'Delivery', 'Takeaway', 'Dine-in'],                         │ │
│ │ │   │   'Accessibility': ['Wheelchair-accessible entrance'],                                                   │ │
│ │ │   │   'Offerings': ['Organic dishes', 'Vegetarian options'],                                                 │ │
│ │ │   │   'Dining options': ['Dessert'],                                                                         │ │
│ │ │   │   'Amenities': ['Good for kids'],                                                                        │ │
│ │ │   │   'Atmosphere': ['Casual'],                                                                              │ │
│ │ │   │   'Crowd': ['Groups'],                                                                                   │ │
│ │ │   │   'Planning': ['Accepts reservations'],                                                                  │ │
│ │ │   │   'Payments': ['Credit cards']                                                                           │ │
│ │ │   }                                                                                                          │ │
│ │ )
Owner
Guilherme Latrova
Sportist, Creator, Software writer, Coffee appreciator, Lucky husband and God servant :)
Guilherme Latrova
抢京东茅台脚本,定时自动触发,自动预约,自动停止

jd_maotai 抢京东茅台脚本,定时自动触发,自动预约,自动停止 小白信用 99.6,暂时还没抢到过,朋友 80 多抢到了一瓶,所以我感觉是跟信用分没啥关系,完全是看运气的。

Aruelius.L 117 Dec 22, 2022
Simple tool to scrape and download cross country ski timings and results from live.skidor.com

LiveSkidorDownload Simple tool to scrape and download cross country ski timings and results from live.skidor.com Usage: Put the python file in a dedic

0 Jan 07, 2022
Scrapy-soccer-games - Scraping information about soccer games from a few websites

scrapy-soccer-games Esse projeto tem por finalidade pegar informação de tabela d

Caio Alves 2 Jul 20, 2022
Poolbooru gelscraper - a simple python script for scraping images off gelbooru pools.

poolbooru_gelscraper a simple python script for scraping images off gelbooru pools. modules required:requests_html, and os by default saves files with

savantshuia 1 Jan 02, 2022
A repository with scraping code and soccer dataset from understat.com.

UNDERSTAT - SHOTS DATASET As many people interested in soccer analytics know, Understat is an amazing source of information. They provide Expected Goa

douglasbc 48 Jan 03, 2023
a high-performance, lightweight and human friendly serving engine for scrapy

a high-performance, lightweight and human friendly serving engine for scrapy

Speakol Ads 30 Mar 01, 2022
京东秒杀商品抢购Python脚本

Jd_Seckill 非常感谢原作者 https://github.com/zhou-xiaojun/jd_mask 提供的代码 也非常感谢 https://github.com/wlwwu/jd_maotai 进行的优化 主要功能 登陆京东商城(www.jd.com) cookies登录 (需要自

Andy Zou 1.5k Jan 03, 2023
A Web Scraper built with beautiful soup, that fetches udemy course information. Get udemy course information and convert it to json, csv or xml file

Udemy Scraper A Web Scraper built with beautiful soup, that fetches udemy course information. Installation Virtual Environment Firstly, it is recommen

Aditya Gupta 15 May 17, 2022
WebScraper - A script that prints out a list of all EXTERNAL references in the HTML response to an HTTP/S request

Project A: WebScraper A script that prints out a list of all EXTERNAL references

2 Apr 26, 2022
A simple django-rest-framework api using web scraping

Apicell You can use this api to search in google, bing, pypi and subscene and get results Method : POST Parameter : query Example import request url =

Hesam N 1 Dec 19, 2021
This code will be able to scrape movies from a movie website and also provide download links to newly uploaded movies.

Movies-Scraper You are probably tired of navigating through a movie website to get the right movie you'd want to watch during the weekend. There may e

1 Jan 31, 2022
A tool can scrape product in aliexpress: Title, Price, and URL Product.

Scrape-Product-Aliexpress A tool can scrape product in aliexpress: Title, Price, and URL Product. Usage: 1. Install Python 3.8 3.9 padahal halaman ins

Rahul Joshua Damanik 1 Dec 30, 2021
This is a web crawler that works on employ email data by gmane.org and visualizes it in different ways.

crawler_to_visual_gmane Analyzing an EMAIL Archive from gmane and vizualizing the data using the D3 JavaScript library. This is a set of tools that al

Saim Zafar 1 Dec 20, 2021
淘宝茅台抢购最新优化版本,淘宝茅台秒杀,优化了茅台抢购线程队列

淘宝茅台抢购最新优化版本,淘宝茅台秒杀,优化了茅台抢购线程队列

MaoTai 118 Dec 16, 2022
Twitter Claimer / Swapper / Turbo - Proxyless - Multithreading

Twitter Turbo / Auto Claimer / Swapper Version: 1.0 Last Update: 01/26/2022 Use this at your own descretion. I've only used this on test accounts and

Underscores 6 May 02, 2022
PaperRobot: a paper crawler that can quickly download numerous papers, facilitating paper studying and management

PaperRobot PaperRobot 是一个论文抓取工具,可以快速批量下载大量论文,方便后期进行持续的论文管理与学习。 PaperRobot通过多个接口抓取论文,目前抓取成功率维持在90%以上。通过配置Config文件,可以抓取任意计算机领域相关会议的论文。 Installation Down

moxiaoxi 47 Nov 23, 2022
Scrapes the Sun Life of Canada Philippines web site for historical prices of their investment funds and then saves them as CSV files.

slocpi-scraper Sun Life of Canada Philippines Inc Investment Funds Scraper Install dependencies pip install -r requirements.txt Usage General format:

Daryl Yu 2 Jan 07, 2022
Using Python and Pushshift.io to Track stocks on the WallStreetBets subreddit

wallstreetbets-tracker Using Python and Pushshift.io to Track stocks on the WallStreetBets subreddit.

91 Dec 08, 2022
A Powerful Spider(Web Crawler) System in Python.

pyspider A Powerful Spider(Web Crawler) System in Python. Write script in Python Powerful WebUI with script editor, task monitor, project manager and

Roy Binux 15.7k Jan 04, 2023
Discord webhook spammer with proxy support and proxy scraper

Discord webhook spammer with proxy support and proxy scraper

3 Feb 27, 2022