辅导案例-JOUR7280-Assignment 1

欢迎使用51辅导,51作业君孵化低价透明的学长辅导平台,服务保持优质,平均费用压低50%以上! 51fudao.top
JOUR7280 Big Data Analytics for Media and
Communication 2019-20 S2
Individual Assignment 1 (20’)
Released at 09 Mar 2020
Deadline: 23 Mar 2020
Overview
Task 1 is an automated online data acquisition challenge, i.e., to scrape some required information
from one or many static webpages with Python Requests and BeautifulSoup packages in an
automated manner, and store the information in a structured, machine readable data format as the
output.
Information to be fetched
(starting from Q2, all the answers should be accompanied by the Python code)
Q1. The problem: what is the webpage (URL) you want to scrape? (1 mark)
suggestion: priority will be given to “public-oriented” webpages, such as governmental officials,
academic institutions, Wikipedia public pages, NGOs (please check their term of use).
Q2. Quick warming up: please figure out the following information from the webpage from your
Python code:
1. What is the “title” of this webpage, as stipulated by the “head” section? (1 mark)
2. What is the “keywords” of this webpage, as stipulated by the “head” section? (hint: = “description”> …,1 mark)
3. What is the “description” of this webpage, as stipulated by the “head” section? (hint: name = “description”> …, 1 mark)
4. please print all the text of the paragraphs of your webpage (1 mark).
Q3. Please identify the elements you want to scrape, by make a screen snapshot of your browser,
with the “development mode.” Hover your mouse above the area of the element you want to scrape
(a rough and approximate area will do). (2 marks)
Q4. Now the scraping. Please scrape the information from the webpage, with the following
requirements:
1. you are expected to scrape a set of information from the web, i.e., elements in the tables,
elements in several tags, several files, or any other information that fit your purposes (6 marks);
2. you are expected to use at least one of the control flow statements to fine tune the code (6
marks);
3. You are expected to store the information in a structured format, and output the file in CSV
format (2 marks). Attention: please use “utf-8” as the encoding.
4. Bonus (5 marks, optional) is offered for scraping multiple pages, using pre-defined rules.
Submission requirements
Please submit the following items:
1. A word file (word, not PDF), spelling out the your answer to Q1 and Q2, for Q3, write the
following information:
problem you define (what is the webpage);
the tasks your code can perform (i.e., “scrape all the temperature information from the
weather service); and
the output (i.e., “the output is a CSV file; each row represents one city, and there are three
columns representing variables: the weather, the temperature, and the degree”…etc).
2. An .ipynb file, including all the codes. You can write comments (#) to briefly spell out the function
of the codes;
3. A CSV file as the output.
4. No required word count – as long as you make sure all necessary information is clearly spelled
out.
5. Once done, please put the three files in a folder, name the foler as this:
“jour7280_assignment01_Name_studentid”
6. Please zip (zip, not RAR) the folder, and submit the zip file to Moodle.
51作业君

Email:51zuoyejun

@gmail.com

添加客服微信: abby12468