Giter Site home page Giter Site logo

farhanibrhim / python_eda Goto Github PK

View Code? Open in Web Editor NEW

This project forked from drshahizan/python_eda

0.0 0.0 0.0 211.16 MB

This topic explains about the implementation of exploratory data analysis (EDA). A total of 21 EDA case studies have been implemented using the Malaysian dataset.

HTML 3.67% Jupyter Notebook 96.33%

python_eda's Introduction

Stars Badge Forks Badge Pull Requests Badge Issues Badge GitHub contributors Visitors

๐ŸŒŸ Hit star button to save this repo in your profile

About Us

The information on this Github is part of the materials for the subject High Performance Data Processing (SECP3133). This folder contains general Exploratory Data Analysis (EDA) information as well as EDA case studies using Malaysian datasets. This case study was created by a Bachelor of Computer Science (Data Engineering), Universiti Teknologi Malaysia student.

Exploratory Data Analysis

Exploratory Data Analysis (EDA) is a crucial step in the data analysis process that involves examining and summarizing a dataset to understand its characteristics, identify patterns, and gain insights into the data. EDA is typically performed before more advanced statistical and machine learning techniques are applied and helps in forming hypotheses, selecting appropriate modeling approaches, and ensuring data quality. Here are some key components and techniques used in EDA:

Exploratory Data Analysis (EDA) is a crucial step in the data analysis process that involves examining and summarizing a dataset to understand its characteristics, identify patterns, and gain insights into the data. EDA is typically performed before more advanced statistical and machine learning techniques are applied and helps in forming hypotheses, selecting appropriate modeling approaches, and ensuring data quality. Here are some key components and techniques used in EDA:

  1. Data Summary: Begin by understanding the basic information about the dataset, such as the number of rows and columns, data types, missing values, and summary statistics (mean, median, standard deviation, etc.).

  2. Data Visualization: Visualizing data through plots and charts can provide a clearer understanding of its distribution and patterns. Common types of visualizations include histograms, box plots, scatter plots, and bar charts.

  3. Data Distribution: Analyze the distribution of variables to determine whether they follow normal, uniform, or other types of distributions. This can impact the choice of statistical tests and modeling techniques.

  4. Correlation Analysis: Explore the relationships between variables using correlation matrices, scatter plots, and other correlation measures. This helps identify potential dependencies and multicollinearity.

  5. Outlier Detection: Identify and handle outliers in the data. Outliers can significantly affect statistical measures and model performance.

  6. Categorical Variables: Examine the distribution of categorical variables through frequency tables, bar plots, and pie charts. This helps understand the composition of categorical data.

  7. Data Transformation: Apply transformations (e.g., log transformation, standardization) to make the data more suitable for analysis, especially if it doesn't meet assumptions of statistical methods.

  8. Feature Engineering: Create new variables or features that might be more informative or relevant for the analysis. This could involve aggregating, combining, or extracting information from existing variables.

  9. Missing Data Handling: Deal with missing data, either by imputing missing values or excluding incomplete records. The choice of method depends on the nature of the data and the problem at hand.

  10. Hypothesis Testing: If relevant, perform hypothesis tests to determine whether observed differences or relationships in the data are statistically significant.

  11. Data Transformation: Consider scaling or encoding categorical variables for modeling. This can include one-hot encoding, label encoding, or other techniques.

  12. Dimensionality Reduction: Use techniques like Principal Component Analysis (PCA) or t-Distributed Stochastic Neighbor Embedding (t-SNE) to reduce the dimensionality of the data while preserving important information.

  13. Time Series Analysis: For time series data, analyze trends, seasonality, and autocorrelation patterns. Techniques like autocorrelation plots and decomposition can be helpful.

  14. Geospatial Analysis: When dealing with geographic data, use maps, geospatial plots, and spatial statistics to understand spatial patterns and relationships.

  15. Text Analysis: If the dataset contains text data, perform text mining and sentiment analysis to extract insights from the textual content.

EDA is an iterative process, and the specific techniques and tools used can vary depending on the nature of the data and the objectives of the analysis. It plays a crucial role in gaining an initial understanding of the data, guiding subsequent analysis, and making informed decisions about the next steps in a data science or analytical project.

Why is EDA so important in data science?

โœ…๏ธ The main purpose of EDA is to help you look at the data before making any assumptions. In addition to better understanding the patterns in the data or detecting unusual events, it also helps you find interesting relationships between variables.

โœ…๏ธ Data scientists can use exploratory analysis to ensure that the results they produce are valid and relevant to desired business outcomes and goals.

โœ…๏ธ EDA also helps stakeholders by verifying that they are asking the right questions.

โœ…๏ธ EDA can help to answer questions about standard deviations, categorical variables, and confidence intervals.

โœ…๏ธ After the exploratory analysis is completed and the predictions are determined, its features can be used for more complex data analysis or modeling, including machine learning.

Python

๐Ÿ‘‰ Python is a popular programming language for data science and has several libraries and tools that are commonly used for EDA such as:

  1. Pandas: a library for data manipulation and analysis.
  2. Numpy: a library for numerical computing in Python.
  3. Scikit-learn: Scikit-learn is a machine learning library, but it also includes tools for data preprocessing, feature selection, and dimensionality reduction, which are essential for EDA.
  4. Matplotlib: a plotting library for creating visualizations.
  5. Seaborn: a library based on matplotlib for creating visualizations with a higher-level interface.
  6. Plotly: an interactive data visualization library.

In EDA, you might perform tasks such as cleaning the data, handling missing values, transforming variables, generating summary statistics, creating visualizations (e.g. histograms, scatter plots, box plots), and identifying outliers. All of these tasks can be done using the above libraries in Python.

๐Ÿ“– Notes

Basic Concept

Code & Practice

Videos

Kaggle: Notebook

Github

No. Repository Name Description
1 PacktPublishing/Hands on Exploratory Data analysis with Python This repository is likely associated with a book or course from Packt Publishing, focusing on hands-on exploratory data analysis with Python. It may contain code examples and materials for learning EDA.
2 code4kunal/eda-python-examples This repository likely contains Python examples and code snippets for exploratory data analysis (EDA). It may serve as a resource for those looking to learn EDA techniques with Python.
3 SouRitra01/Exploratory-Data-Analysis-EDA-in-Banking-Using-Python This repository appears to be focused on conducting exploratory data analysis (EDA) in the context of banking using Python. It may contain datasets and code for EDA in the banking domain.
4 sandipanpaul21/EDA-in-Python This repository is likely dedicated to exploratory data analysis (EDA) in Python. It may contain Python scripts, Jupyter notebooks, and related materials for EDA projects.
5 vharivinay/python-eda-viz This repository may be focused on Python-based exploratory data analysis and data visualization. It could provide code and examples for creating data visualizations during EDA.
6 demonpratapdemon/Exploratory-Data-Analysis-EDA-and-PreProcessing This repository seems to cover both exploratory data analysis (EDA) and data preprocessing in Python. It may contain code and resources for these data preparation tasks.
7 PacktPublishing/Python-for-Data-Analysis-step-by-step-with-projects- This repository is likely associated with a book or course from Packt Publishing, focusing on Python for data analysis with step-by-step projects. It may include code and project materials.
8 sandyy2505/Cardio Good Fitness Project This repository may contain code and data related to a fitness project, possibly involving data analysis and visualization in the context of cardio fitness.
9 ajaymache/Data analysis of used car database This repository is likely focused on data analysis of a used car database. It may provide Python code and data for analyzing and exploring information related to used cars.

๐Ÿ“– Lab

No Dataset Colab GitHub
1 Boston Open in Colab Open in GitHub
2 Car Features and MSRP Open in Colab Open in GitHub
3 Housing Dataset Open in Colab Open in GitHub
4 United Nations Development Corporation Open in Colab Open in GitHub

The provided list comprises a collection of case studies, each with a title and accessibility information on platforms like Colab and GitHub. These case studies likely involve data analysis and exploration. For instance, "404 Error" may involve exploring property-related data in Kuala Lumpur, while "Alrite" could be centered around the exportation of plantation products in Sarawak. "BEFE" appears to focus on COVID-19 clusters in Malaysia, "Boboiboy" on property listings in Kuala Lumpur, and "COLBY" on the results of the 14th General Election in Malaysia. "FANTOM" likely tracks daily recorded COVID-19 cases at the state level, "HAHA" pertains to foreign direct investment in Malaysia, and "HD" may involve land usage analysis in Tampin for 2021. Other case studies cover topics such as elections, healthcare, real estate, population, and more, providing a diverse range of data exploration possibilities.

Automated EDA Tools

EDA is a vital but time-consuming task in a data project. Here are 10 open-source tools that generate an EDA report in seconds.

Library Description Web Github
SweetViz - In-depth EDA report in two lines of code.
- Covers information about missing values, data statistics, etc.
- Creates a variety of data visualizations.
- Integrates with Jupyter Notebook.
๐ŸŒ :octocat:
Pandas-Profiling - Generate a high-level EDA report of your data in no time.
- Covers info about missing values, data statistics, correlation etc.
- Produces data alerts.
- Plots data feature interactions.
๐ŸŒ :octocat:
DataPrep - Supports Pandas and Dask DataFrames.
- Interactive Visualizations.
- 10x Faster than Pandas based tools.
- Covers info about missing values, data statistics, correlation etc.
- Plots data feature interactions.
๐ŸŒ :octocat:
AutoViz - Supports CSV, TXT, and JSON.
- Interactive Bokeh charts.
- Covers info about missing values, data statistics, correlation etc.
- Presents data cleaning suggestions.
๐ŸŒ :octocat:
D-Tale - Runs common Pandas operation with no-code.
- Exports code of analysis.
- Covers info about missing values, data statistics, correlation etc.
- Highlights duplicates, outliers, etc.
- Integrates with Jupyter Notebook.
๐ŸŒ :octocat:
dabl - Primarily provides visualizations.
- Covers wide range of plots: Scatter pair plots. Histograms.
- Target distribution.
๐ŸŒ :octocat:
QuickDA - Get overview report of dataset.
- Covers info about missing values, data statistics, correlation etc.
- Produces data alerts.
- Plots data feature interactions.
๐ŸŒ :octocat:
Datatile - Extends Pandas describe().
- Provides column stats: column type count, missing, column datatype.
- Mostly statistical information.
๐ŸŒ :octocat:
Lux - Provides visualization recommendations.
- Supports EDA on a subset of columns.
- Integrates with Jupyter Notebook.
- Exports code of analysis.
๐ŸŒ :octocat:
ExploriPy - Performs statistical testing.
- Column type-wise distribution: Continuous, Categorical
- Covers info about missing values, data statistics, correlation etc.
๐ŸŒ :octocat:

Big Data refers to the vast and complex datasets that exceed the capabilities of traditional data processing tools. It is characterized by the three V's: volume (large amounts of data), velocity (rapid data generation), and variety (different data types). Exploratory Data Analysis (EDA), on the other hand, is a data analysis approach that involves summarizing, visualizing, and understanding the key characteristics of a dataset to uncover insights and patterns. In the context of Big Data, EDA plays a crucial role in making the data more manageable by identifying relevant subsets, trends, and anomalies, enabling data scientists to extract meaningful information and inform decision-making processes.

Contribution ๐Ÿ› ๏ธ

Please create an Issue for any improvements, suggestions or errors in the content.

You can also contact me using Linkedin for any other queries or feedback.

Visitors

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.