Mainly memos for myself.
Create a scraping tool to get pachinko data
It is a program to get the number of G and the number of big hits. This is almost complete, and we have been able to acquire 5 units in the test, so we should be able to acquire all of them without any problems. Click here for the acquisition flow
I also tried to get it with read_html in the table, but when joining each data frame, I could not join well, so I got only the information I wanted in a list and converted and joined them into a data frame. ..
After that, adjust the type of the acquired data.
It is a program to acquire the slump graph of each unit.
Click here for the acquisition flow you are thinking about
What you have to pay attention to in the image is that the SRC is partly a relative path, whether the site is taking measures for the data of the day. I can't feel the regularity of which model has a relative path. Therefore, the data to be acquired is basically the data of the previous day. It is necessary to investigate what time the site will switch.
Next, it is a program that analyzes the image of the slump graph and converts it into data.
Click here for the flow of thinking
Especially because it's close to my own memo, I don't think it will be helpful to anyone.
Recommended Posts