"en"/> Data Mining (2017-2018): Practicals 1: John Samuel
Creative Commons License

Goals

Exercises

  1. Installation and setting up of pip, scikit-learn and jupyter
  2. Parsing and reading CSV/TSV files
  3. Parsing and reading JSON files
  4. Querying external data sources (Query endpoints and API)
  5. Performing classical data analyses

Exercise 1.1

Installation

Run the following commands on your virtual machines

pip

$ sudo apt install python-pip

jupyter

$ python3 -m pip install jupyter

scikit-learn

$ python3 -m pip install scikit-learn

numpy

$ python3 -m pip install numpy

pandas

$ python3 -m pip install pandas

matplotlib

$ python3 -m pip install matplotlib

Hello World!

If your installation is successful, you can run

$ mkdir TP1 && cd TP1 $ jupyter notebook

A new page will appear on your browser and you will see the following image

Click on the tab 'Running'. You see no notebooks running (if it's the first time you are running jupyter).

Go back to the 'Files' tab and click on New and choose Python3 under Notebook

A new tab will open as shown below. Write the following code in the cell

print("Hello World!")

You can go to any cell and press 'Run'

By default, your Notebook is named 'Untitled'. You can rename it as shown below, by clicking on the name 'Untitled' and by giving a new name.

Now go back to the 'Files' tab and you can see the renamed notebook. You can click on your notebook at any time to continue working.

Now let's continue working on your current Notebook. Write the following code to check whether scikit is properly installed.

The code below shows the available datasets from scikit.

from sklearn import datasets
print(datasets.__all__)

Now, you are ready to go!!

Note:, You can stop Jupyter notebook at any point by typing "Ctrl+c" on the terminal and by pressing 'y' to confirm shutdown.

Exercise 1.2

Most of the time, we work with CSV (comma-separated values) files for data analysis. A CSV file consits of one or more lines and each line has one or more values separated by commas. One can consider every line as a row and every value in a row as a column value. The first row is sometimes used to describe the column names.

Copy the file pl.csv to your current working directory (where you are running Jupyter: TP1) and use the following code to parse the csv file. Note the column names and datatypes (U100, i4)

import numpy as np
dataset = np.loadtxt("pl.csv", dtype={'names': ('name', 'year'), 'formats': ('U100', 'i4')},
   skiprows=1, delimiter=",", encoding="UTF-8")
print(dataset)

CSV support in numpy is different from Python's default CSV reader because of its capability to support the data types. Before continuing, take a deep look at numpy.loadtxt.

Copy the file pl.tsv to your current working directory and use the following code to parse the tsv file.

import numpy as np
dataset = np.loadtxt("pl.tsv", dtype={'names': ('name', 'year'), 'formats': ('U100', 'i4')},
   skiprows=1, delimiter="\t", encoding="UTF-8")
print(dataset)

Note the changes in the above code compared to the previous one. A TSV file is a tab-separated file, i.e., the column values are separated by a tab (\t).

Exercise 1.3

Most of the external data sources may provide their data in JSON format. Our next exercise is to parse JSON files. Copy the file pl.json to your current working directory and use the following code to parse the JSON file. In this exercise, we use Pandas python package to parse the JSON file to obtain a Pandas DataFrame. Try using methods like transpose, count etc.

from pandas.io.json import json_normalize
import pandas as pd
import json

data = json.load(open('pl.json'))
dataframe = json_normalize(data)
print(dataframe)

Exercise 1.4

In this exercise, we will take a look at how to download data from external data sources using special query interfaces. Take for example, above data was obtained from Wikidata query interface with the query examples. See the screenshot given below.

Given below is the code to read data from an external data source. Use this url: https://query.wikidata.org/sparql?query=SELECT%20%3FlanguageLabel%20(YEAR(%3Finception)%20as%20%3Fyear)%0AWHERE%0A%7B%0A%20%20%23instances%20of%20programming%20language%0A%20%20%3Flanguage%20wdt%3AP31%20wd%3AQ9143%3B%0A%20%20%20%20%20%20%20%20%20%20%20%20wdt%3AP571%20%3Finception%3B%0A%20%20%20%20%20%20%20%20%20%20%20%20rdfs%3Alabel%20%3FlanguageLabel.%0A%20%20FILTER(lang(%3FlanguageLabel)%20%3D%20%22en%22)%0A%7D%0AORDER%20BY%20%3Fyear%0ALIMIT%20100&format=json and replace ".." by the previous URL value.

import urllib.request
import json
import pandas as pd

url = "..."
response = urllib.request.urlopen(url)
responsedata = json.loads(response.read().decode('utf-8'))

array = []

for data in responsedata['results']['bindings']:
  array.append([data['year']['value'], data['languageLabel']['value']])
dataframe = pd.DataFrame(array, columns=['year', 'languageLabel'])
dataframe = dataframe.astype(dtype= {"year":"<i4", "languageLabel":"<U200"})
print(dataframe)

Exercise 1.5

This final exercise will use some basic data analyses. Continuing with the code in Exercise 1.4, le's count the number of programming languages released in a year.

grouped = dataframe.groupby('year').count()
print(grouped)

You can also use multiple aggregate functions using agg()

grouped = dataframe.groupby('year').agg(['count'])
print(grouped)

Till now, we worked with tables having two columns. Now we focus on tables with three columns (programming language, year, paradigm). Copy the file plparadigm.json to your working directory. And test the following program.

from pandas.io.json import json_normalize
import pandas as pd
import json

jsondata = json.load(open('plparadigm.json'))
array = []

for data in jsondata:
  array.append([data['year'], data['languageLabel'], data['paradigmLabel']])
dataframe = pd.DataFrame(array, columns=['year', 'languageLabel', 'paradigmLabel'])
dataframe = dataframe.astype(dtype= {"year" : "int64", "languageLabel" : "<U200", "paradigmLabel" : "<U200"})

grouped = dataframe.groupby(['year', 'paradigmLabel']).agg(['count'])
print(grouped)

Now test the following program. Compare the difference in output.

grouped = dataframe.groupby(['paradigmLabel', 'year']).agg(['count'])
print(grouped)

Your next goal is to run the following query to get the population information of different countries (limited to 1000 rows). Run the following query on Wikidata query service and download the JSON file.

SELECT DISTINCT ?countryLabel (YEAR(?date) as ?year) ?population
WHERE {
 ?country wdt:P31 wd:Q6256; #Country
   p:P1082 ?populationStatement;
  rdfs:label ?countryLabel. #Label
 ?populationStatement ps:P1082 ?population; #population
  pq:P585 ?date. #period in time
 FILTER(lang(?countryLabel)="en") #Label in English
}
ORDER by ?countryLabel ?year
LIMIT 1000

Now, compute and display the following information (using various operations available in pandas library):
  1. The population of countries in alphabetical order of their names and ascending order of year.
  2. The latest available population of every country
  3. The country with the lowest and highest population (considering the latest population)

Your next goal is to run the following query to get information related to scientific articles published after 2010 (limited to 1000 rows). Run the following query on Wikidata query service and download the JSON file. It gives you the following information related to the scientific article: title, main subject and publication year.

SELECT ?title ?subjectLabel ?year
{
  ?article wdt:P31 wd:Q13442814; #scientific article
   wdt:P1476 ?title; #title of the article
   wdt:P921 ?subject; #main subject
   wdt:P577 ?date. #publication date
  ?subject rdfs:label ?subjectLabel.
  BIND(YEAR(?date) as ?year).
  #published after 2010
  FILTER(lang(?title)="en" && lang(?subjectLabel)="en" && ?year>2010)
}
LIMIT 1000

Now, compute and display the following information (using various operations available in pandas library):

  1. The number of articles published on different subjects every year.
  2. Top subject of interest to the scientific community every year(based on the above query results).
  3. Top 10 subjects of interest to the scientific community (based on the above query results) since 2010.

Submission

References

Link