This week topic's is Data Scraping. We lerned to use multiple tools for the task such as...
Sadly, to use these tools in their full potential (except the Python one) you have to pay for unlocking them. But luckily, the course only needs a bare-minimum of result; just to test if I understand how to scarpe data or not.
I found OutWit was the most difficult to use, since the program needed a basic knowledge in coding. However, it wasn't too hard to understand when follow the indtruction carefully. But this was why I didn't conduct this experiment with other web than Reddit.
With WebScarper.io I can "point and click" what data I wanted to scrape. The process of choosing and naming data type could be tedious, but it's easier than reading a code. I got some headlines and links from CNN with this method.
ScrapeHero is pretty simple, but only allowed 25 page requests. Moreover, I found that if the data is too big, it would take forever to processing. So I only scrape some comments from Amazon and called it a day.
Last but not least, the Python. It's overwhelming at first, but Ben helped us a lot. I have to admit that I still confused on some part, and have to work on how to understanding the method. Although the instruction said I don't need to, I think it's fun to know many coding languages.