Giter Site home page Giter Site logo

annajiat / introduction-to-ml-with-sklearn Goto Github PK

View Code? Open in Web Editor NEW

This project forked from kaust-vislab/introduction-to-ml-with-sklearn

0.0 1.0 0.0 2.63 MB

Introduction to Machine Learning with Scikit-Learn workshop materials

License: BSD 3-Clause "New" or "Revised" License

Shell 0.45% Dockerfile 2.48% Jupyter Notebook 97.07%

introduction-to-ml-with-sklearn's Introduction

badge badge

Introduction to Machine Learning with Scikit Learn Workshop

The KAUST Visualization Core Lab (KVL) will host an Introduction to Machine Learning with Scikit-Learn workshop on Tuesday, 30 March 2021 from 1 to 5 pm AST. The workshops will largely follow Chapter 2 of Hands-on Machine Learning with Scikit-Learn, Keras, and TensorFlow which walks through the process of developing an end-to-end machine learning project with Scikit-Learn.

  • Working with Real Data
  • Understanding the Big Picture
  • Getting the Data
  • Discovering and Visualizing the Data to Gain Insights
  • Preparing the Data for Machine Learning Algorithms
  • Selecting and Training a Model
  • Fine Tuning Your Model

This hands-on lesson is part of the Introduction to Data Science Workshop Series being offered by KVL as part of our on-going efforts to build capacity in core data science skills both at KAUST and within the Kingdom of Saudi Arabia (KSA).

The workshop will be hosted on Zoom with the relevant details sent out to registered participants prior to the event. The workshop will be recorded with the recording released via the KVL YouTube Channel.

Workshop registration

If you are interested in registering for this workshop, then please see the following links.

KAUST Affiliated

Non-KAUST Affiliated

Project organization

Project organization is based on ideas from Good Enough Practices for Scientific Computing.

  1. Put each project in its own directory, which is named after the project.
  2. Put external scripts or compiled programs in the bin directory.
  3. Put raw data and metadata in a data directory.
  4. Put text documents associated with the project in the doc directory.
  5. Put all Docker related files in the docker directory.
  6. Install the Conda environment into an env directory.
  7. Put all notebooks in the notebooks directory.
  8. Put files generated during cleanup and analysis in a results directory.
  9. Put project source code in the src directory.
  10. Name all files to reflect their content or function.

Using Conda

Creating the Conda environment

After adding any necessary dependencies for your project to the Conda environment.yml file (or the requirements.txt file), you can create the environment in a sub-directory of your project directory by running the following command.

ENV_PREFIX=$PWD/env
conda env create --prefix $ENV_PREFIX --file environment.yml --force

Once the new environment has been created you can activate the environment with the following command.

conda activate $ENV_PREFIX

Note that the ENV_PREFIX directory is not under version control as it can always be re-created as necessary.

If you wish to use any JupyterLab extensions included in the environment.yml and requirements.txt files then you need to activate the environment and rebuild the JupyterLab application using the following commands to source the postBuild script.

conda activate $ENV_PREFIX # optional if environment already active
source postBuild

For your convenience these commands have been combined in a shell script ./bin/create-conda-env.sh. Running the shell script will create the Conda environment, activate the Conda environment, and build JupyterLab with any additional extensions. The script should be run from the project root directory as follows.

./bin/create-conda-env.sh

Listing the full contents of the Conda environment

The list of explicit dependencies for the project are listed in the environment.yml file. To see the full lost of packages installed into the environment run the following command.

conda list --prefix $ENV_PREFIX

Updating the Conda environment

If you add (remove) dependencies to (from) the environment.yml file or the requirements.txt file after the environment has already been created, then you can re-create the environment with the following command.

conda env create --prefix $ENV_PREFIX --file environment.yml --force

If you add any additional JupyterLab extensions to the postBuild file, then you should run the environment creation script again.

./bin/create-conda-env.sh

Using Docker

In order to build Docker images for your project and run containers you will need to install Docker and Docker Compose.

Detailed instructions for using Docker to build and image and launch containers can be found in the docker/README.md.

introduction-to-ml-with-sklearn's People

Contributors

davidrpugh avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.