Giter Site home page Giter Site logo

mbar0075 / foundations-of-data-science Goto Github PK

View Code? Open in Web Editor NEW
0.0 1.0 0.0 7.46 MB

Deliverables relating to the Foundations of Data Science University Unit

License: MIT License

HTML 100.00%
data-analysis data-cleaning data-science probability statistics bayes-rule probability-theory r statistical-analysis distribution

foundations-of-data-science's Introduction

Foundations of Data Science

Author

Matthias Bartolo 0436103L

Preview:


Description of Task:

The aim of these projects was to develop skills required to navigate an increasingly data-driven world. The focus was on exploring different data science platforms and utilizing the R programming language. The projects aimed to provide a comprehensive understanding of various techniques related to data collection, storage, access, cleaning, transformation, and data management. Additionally, an emphasis was placed on data analytics using probability theory, distributions, sampling techniques, multivariate thinking, correlation and causation, and hypothesis testing. The projects also involved gaining familiarity with visualization tools and libraries to create visually appealing and informative visualizations and dashboards.


Group Project:

This project revolved around a dataset from Rent-Apart, a Dutch rental-aggregation company. The dataset encompassed house prices, rental agencies, cities, bedrooms, and surface area. Thisi project was structured into tasks that facilitated a deeper comprehension of the data, data cleaning, data analysis, and interpretation and presentation of the findings.

The initial task involved an in-depth analysis of the dataset to extract meaningful insights. Statistical analysis techniques were applied to enhance the understanding of the dataset. This preliminary exploration shed light on the distribution of the features and guided subsequent analysis decisions.

Data cleaning was a pivotal step in ensuring the accuracy and reliability of the analysis. Null values in the dataset were addressed, taking into consideration specific cities and utilizing surface area and number of bedrooms to impute missing values. The handling of instances where no data was available for a city was approached thoughtfully.

APPSM <- vector()

for (i in unique(dataInitial$city)){#Looping through all the unique cities and 
  #calculating the mean of the house prices and surface area
  tmp <- which(dataInitial$city==i)

  Hprice <- c(dataInitial[tmp,1])#Storing the respective data in vectors
  Surface <- c(dataInitial[tmp,5])

  MP <- mean(Hprice)#Working out the mean
  MS <- mean(Surface)

  APPSM <- append(APPSM,(MP/MS)) #Working out and storing the 
  #average Price Per Square Meter
}

cities <- unique(dataInitial$city)

loc <- cities[which(APPSM==max(APPSM))]#Retrieving the city name through the 
#max and match functions
print(loc)#Printing the respective location of the max value
print(max(APPSM))#Printing the maximum value

loc2 <- cities[which(APPSM==min(APPSM))]#Retrieving the city name through the 
#min and match functions
print(loc2)#Printing the respective location of the min value
print(min(APPSM))#Printing the minimum value

Building upon the initial exploration and the cleaned dataset, a comprehensive data analysis was conducted. This encompassed calculating sample mean values for price, bedrooms, and floor area across the country, identifying the cities with the highest and lowest living costs in terms of surface area, visualizing the price distribution across the country using heatmaps and maps, examining the distribution and standard deviation of prices in major cities such as Amsterdam and Rotterdam, exploring correlations between variables, and employing regression techniques to predict apartment characteristics and monthly rent in specific cities.

To effectively communicate the findings, a simple static flexdashboard was developed. This served as a concise and informative platform for presenting the analysis from the projects. Additionally, all work was thoroughly documented in an R markdown (.Rmd) document, providing detailed explanations and justifications for the decisions made throughout the projects.


Individual Project:

In this project, a series of challenges were presented, each requiring the application of probability theory and statistical analysis. The project involved exploring different scenarios and calculating probabilities based on given information.

The first challenge, known as the "Queen's Gambit", involved a chess player named Beth playing two games against an opponent of unknown skill level. The opponent could be a beginner, intermediate, or a master, each with different probabilities of winning against Beth. The goal was to determine the probability of Beth winning the first game and, given that she won the first game, the probability of her winning the second game.

In the second challenge, called "Conditionitis", a screening test for a fictitious disease was analyzed. The test had a certain accuracy in identifying the disease and a known rate of false positives. The objective was to calculate the probability of a man having the disease given a positive test result, as well as the probability of having the disease given a negative test result.


The third challenge, referred to as "Bottled Water", focused on assessing whether a water dispensing machine was operating properly based on the amount of water dispensed into bottles. A sample of 35 bottles was taken, and the average amount dispensed and its variance were provided. The task was to determine if there was evidence that the machine needed repairs.

In each of these challenges, hypotheses were formulated to guide the analysis. Critical values and rejection regions were established to make decisions based on the calculated test statistics. Sample parameters were calculated and compared to the critical values to draw conclusions.

Closing Remarks:

These projects provided valuable hands-on experience in applying data science techniques, analyzing real-world datasets, and presenting findings effectively. The skills and knowledge acquired through these projects are essential for navigating the data-driven landscape and making meaningful contributions in the field of data science.

Deliverables:

The repository includes two directories:

  1. Data_Science_Individual: Individual Project utilising R
  2. GroupProject: Group Project utilising R

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.