当前位置:网站首页>Entrepreneurship is a little risky. Read the data and do a business analysis
Entrepreneurship is a little risky. Read the data and do a business analysis
2022-07-02 00:52:00 【Coriander Chat Game】
1、 cause
After the new year , As a program is also very old , So I want to do a sideline , But blind entrepreneurship is not enough , Mr. Lu Xun said : Direction over effort , So choosing the right direction is very important , The first step of data research , Don't act blindly , A waste of time , Waste energy , The main waste of my hard-earned money .
I usually order takeout during working hours , So I want to make takeout , Because I am Xiaobai , So I want to find a brand takeout store to join , But most of them know that franchise stores are relatively stupid , So be cautious , Do research .
2、 Data crawling
The platforms selected for data crawling are two big takeout platforms , Here, choose one of them for analysis. I want to see the take away order volume of the business district , Do a survey , The data content crawled is the category of takeout , A single quantity , And the name , Location , These basic data , And then make a summary , Analyze the data in the business district .
3、 Crawling step
1、 Confirm the URL you crawled
The data source of crawling is a app, The local solutions are nocturnal simulator and Charles, Installed the environment and did a few things , I didn't figure it out , This plan was abandoned

stay xxx After a search on our official website, we found h5 The interface of , It seems that this interface is not public , After searching for a long time, I finally found the entrance
After twists and turns, I finally found , But after logging in, you still have to jump to the official website , So enter this URL again
be based on csdn The rules of ,xxx Please replace

The final crawling data interface is :
2, Analyze the website
The data request method is get request , So all the parameters are url in
latitude=31.296829
&longitude=120.736135
&offset=16
&limit=8
&extras[]=activities
&extras[]=tags
&extra_filters=home
&terminal=h5
The first two data are longitude and latitude , That is, the offset of positioning information , This data determines the data pulled , This is where I want to locate .
offset I don't know what this offset means , I guess it may be the number , Is not important
limit I guess how many data per page
terminal Is the type of terminal , Basically
3、 Analyze the returned data
{
has_next: true,
items: - [
- {
restaurant: + {... }
},
- {
restaurant: + {... }
},
- {
restaurant: + {... }
},
- {
restaurant: + {... }
},
- {
restaurant: + {... }
},
- {
restaurant: + {... }
},
- {
restaurant: + {... }
},
- {
restaurant: + {... }
}
],
meta: - {
rankId: "",
rankType: - {
505517688: "33"
}
}
}The most external data format is like this , What we want is restaurant This layer of data , It is also important to analyze
restaurant: - {
act_tag: 0,
activities:[],
address: null,
authentic_id: 502101541,
average_cost: null,
baidu_id: null,
bidding: null,
brand_id: 710858,
business_info: "{"pickup_scheme":"https://tb.xxx.me/wow/a/act/eleme/dailygroup/682/wupr?wh_pid=daily-186737&id=E14670700902593244","ad_info":{"isAd":"false"},"recent_order_num_display":" On sale 1244"}",
closing_count_down: 14234,
delivery_fee_discount: 0,
delivery_mode: + {... },
description: null,
distance: 2804,
favor_time: null,
favored: false,
flavors: + [... ],
float_delivery_fee: null,
float_minimum_order_amount: 20,
folding_restaurant_brand: null,
folding_restaurants:[],
has_story: false,
id: "E14670700902593244",
image_path: "https://img.alicdn.com/imgextra/i2/2212739556234/O1CN01jzFa3R1vvDoDnsojl_!!2212739556234-0-koubei.jpg",
is_new: false,
is_premium: true,
is_star: false,
is_stock_empty: 0,
is_valid: null,
latitude: null,
longitude: null,
max_applied_quantity_per_order: -1,
name: " Fruit cutter ( East Lake CBD shop )",
next_business_time: " Tomorrow, 9:30",
only_use_poi: null,
opening_hours: - [
"9:30/0:10"
],
order_lead_time: 42,
out_of_range: false,
phone: null,
piecewise_agent_fee: + {... },
platform: 0,
posters:[],
promotion_info: null,
rating: 4.7,
rating_count: null,
recent_order_num: 1244,
recommend: + {... },
recommend_reasons: + [... ],
regular_customer_count: 0,
restaurant_info: null,
scheme: "https://h5.xxx.me/newretail/p/shop/?store_id=546110047&geolat=31.296829&geolng=120.736135&o2o_extra_param=%7B%22rank_id%22%3A%22%22%7D",
status: 1,
support_tags: + [... ],
supports:[],
target_tag_path: "35a1bb9025ab98c28112d82f83f73d7ejpeg",
theme: null,
type: 1
}The data seems easy to understand , The data we need here is
name The name of the take out shop
business_info.recent_order_num_display Sales figures
activities Store activity data
opening_hours It's the business hours of the store
support_tags It is the classification of stores
4、 Crawl data
I haven't written all the code for crawling data , So the above analysis
The core of crawling data is pretending to be a normal request , We have seen the data web Interface , Basically nothing
In general, that is :
cookie requirement
js encryption
token Request and so on
Because I don't use it in batches , So it can be web After the client logs in , Use data directly , Disguised as a web That's all right. , At its worst, it can be used directly seleunim To climb , So it's not a big problem
5、 summary
Looking at the data, it seems that some brands have good sales , But I can't see the opening time of the store , So there is no way to judge whether it is a buying behavior , But the average sales volume is not high , A little polarized , The situation of taking out is grim , Earn hard money , The platform also draws from the turnover 15% , Remove the franchise fee , Shop rentals , There are also various equipment costs , It's really a lot of money . In the end, all the money was earned by the platform , As a business, there is basically no profit . therefore Take out is not recommended , Feel stifled .
边栏推荐
- Leetcode skimming: binary tree 03 (post order traversal of binary tree)
- Node -- egg implements the interface of uploading files
- What skills does an excellent software tester need to master?
- Zak's latest "neural information transmission", with slides and videos
- 2022 pinduoduo details / pinduoduo product details / pinduoduo SKU details
- 【八大排序②】选择排序(选择排序,堆排序)
- RFID makes the inventory of fixed assets faster and more accurate
- 【CTF】bjdctf_2020_babystack2
- Qt5.12.9 migration tutorial based on Quanzhi H3
- New version of free mobile phone, PC, tablet, notebook four terminal Website thumbnail display diagram online one click to generate website source code
猜你喜欢

New version of free mobile phone, PC, tablet, notebook four terminal Website thumbnail display diagram online one click to generate website source code

Evolution of Himalayan self-developed gateway architecture

Leetcode skimming: stack and queue 05 (inverse Polish expression evaluation)

excel查找与引用函数

数据库--SqlServer详解

Collection: comprehensive summary of storage knowledge

Geek DIY open source solution sharing - digital amplitude frequency equalization power amplifier design (practical embedded electronic design works, comprehensive practice of software and hardware)

【微信授权登录】uniapp开发小程序,实现获取微信授权登录功能

一名优秀的软件测试人员,需要掌握哪些技能?

【八大排序④】归并排序、不基于比较的排序(计数排序、基数排序、桶排序)
随机推荐
Node - generate wechat permission verification configuration
Take the enclave Park as a sample to see how Yuhua and Shaoshan play the song of Chang Zhu Tan integrated development
SSO single sign on implementation.
Upgraded wechat tool applet source code for mobile phone detection - supports a variety of main traffic modes
Tensorflow tensor convolution, input and convolution kernel dimension understanding
股票开户哪个证券公司比较安全
Collection: comprehensive summary of storage knowledge
How to reflect and solve the problem of bird flight? Why are planes afraid of birds?
How to type spaces in latex
【底部弹出-选择器】uniapp Picker组件——底部弹起的滚动选择器
Leetcode skimming: stack and queue 01 (realizing queue with stack)
Promise和模块块化编程
Node——生成微信权限验证配置
cookie、session、tooken
AIX存储管理之总结篇
Viewing and modifying volume group attributes of Aix storage management (II)
excel数据透视表
Some understandings of graph convolution neural network r-gcn considering relations and some explanations of DGL official code
sso单点登录的实现。
Zak's latest "neural information transmission", with slides and videos