Code Monkey home page Code Monkey logo

datacanvasio / hypernets Goto Github PK

View Code? Open in Web Editor NEW
260.0 19.0 40.0 18.18 MB

A General Automated Machine Learning framework to simplify the development of End-to-end AutoML toolkits in specific domains.

Home Page: https://hypernets.readthedocs.io/

License: Apache License 2.0

Python 99.82% Jupyter Notebook 0.18%
neural-architecture-search hyperparameter-optimization hyperparameter-tuning evolutionary-algorithms monte-carlo-tree-search automl autodl reinforcement-learning mcts nas

hypernets's Introduction

Python Versions Downloads PyPI Version

We Are Hiring!

Dear folks, we are offering challenging opportunities located in Beijing for both professionals and students who are keen on AutoML/NAS. Come be a part of DataCanvas! Please send your CV to [email protected]. (Application deadline: TBD.)

Hypernets: A General Automated Machine Learning Framework

Hypernets is a general AutoML framework, based on which it can implement automatic optimization tools for various machine learning frameworks and libraries, including deep learning frameworks such as tensorflow, keras, pytorch, and machine learning libraries like sklearn, lightgbm, xgboost, etc. It also adopted various state-of-the-art optimization algorithms, including but not limited to evolution algorithm, monte carlo tree search for single objective optimization and multi-objective optimization algorithms such as MOEA/D,NSGA-II,R-NSGA-II. We introduced an abstract search space representation, taking into account the requirements of hyperparameter optimization and neural architecture search(NAS), making Hypernets a general framework that can adapt to various automated machine learning needs. As an abstraction computing layer, tabular toolbox, has successfully implemented in various tabular data types: pandas, dask, cudf, etc.

Overview

Conceptual Model

Illustration of the Search Space

What's NEW !

Installation

Conda

Install Hypernets with conda from the channel conda-forge:

conda install -c conda-forge hypernets

Pip

Install Hypernets with different options:

  • Typical installation:
pip install hypernets
  • To run Hypernets in JupyterLab/Jupyter notebook, install with command:
pip install hypernets[notebook]
  • To run Hypernets in distributed Dask cluster, install with command:
pip install hypernets[dask]
  • To support dataset with simplified Chinese in feature generation,
    • Install jieba package before running Hypernets.
    • OR install Hypernets with command:
pip install hypernets[zhcn]
  • Install all above with one command:
pip install hypernets[all]

To Verify your installation:

python -m hypernets.examples.smoke_testing

Related Links

Documents

Neural Architecture Search

Hypernets related projects

  • Hypernets: A general automated machine learning (AutoML) framework.
  • HyperGBM: A full pipeline AutoML tool integrated various GBM models.
  • HyperDT/DeepTables: An AutoDL tool for tabular data.
  • HyperTS: A full pipeline AutoML&AutoDL tool for time series datasets.
  • HyperKeras: An AutoDL tool for Neural Architecture Search and Hyperparameter Optimization on Tensorflow and Keras.
  • HyperBoard: A visualization tool for Hypernets.
  • Cooka: Lightweight interactive AutoML system.

DataCanvas AutoML Toolkit

Citation

If you use Hypernets in your research, please cite us as follows:

Jian Yang, Xuefeng Li, Haifeng Wu. Hypernets: A General Automated Machine Learning Framework. https://github.com/DataCanvasIO/Hypernets. 2020. Version 0.2.x.

BibTex:

@misc{hypernets,
  author={Jian Yang, Xuefeng Li, Haifeng Wu},
  title={{Hypernets}: { A General Automated Machine Learning Framework}},
  howpublished={https://github.com/DataCanvasIO/Hypernets},
  note={Version 0.2.x},
  year={2020}
}

DataCanvas

Hypernets is an open source project created by DataCanvas.

hypernets's People

Contributors

dc-aps avatar dependabot[bot] avatar enpen avatar jackguagua avatar liuzhaohan0 avatar lixfz avatar oaksharks avatar wyq-1997 avatar zhangxjohn avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

hypernets's Issues

scorer & calc_score

Please make sure that this is a feature request.

System information

  • Hypernets version (you are using):
  • Are you willing to contribute it (Yes/No):

Describe the feature and the current behavior/state.
I found that the user-defined evaluation metric function could not be set in the hypernets experiment, and calc_score() contained few metrics, such as no mape.

Will this change the current api? How?

Any Other info.

There may be some issues with the code

I noticed that using the pip install hyperkeras command did not download the hyperkeras related packages (i used python3.8), so I directly downloaded and referenced the hyperkeras code library, and then used the code about defined in the CNN neural architecture search in the document you provided. However, I found the following error, which part of the definition is missing. I hope to receive your reply as soon as possible. Thank you
%U~DPG%RQ_U7M4_B9I7A7
{8O71K VABXIJLD0A~R%(L8

dataset about dsutils

an error about "from hypernets.frameworks.ml.datasets import dsutils", I think it should be "from deeptables.datasets import dsutils".

A suggestion for GreedyEnsemble

On line 107 of voting's code, i.e.
if self.ensemble_size <= 0: size = predictions.shape[1]

Is it possible to replace it with self.ensemble_size <= 0 or predictions.shape[1] < self.ensemble_size: size = predictions.shape[1]?

independent experiment visualization

  • integration of the dataset page, training page and experiment configuration page
  • front-end: develop and test in the non-Notebook environment
  • back-end: share the js program with Notebook
  • unit test

Hope for references

Thank you very much for your work! Would you please give some references for some of these strategies, for example,Hypernets/hypernets/searchers/evolution_searcher.py: mutate

window pip fail.

ERROR: Complete output from command 'D:\Anaconda\Anaconda3\python.exe' -u -c 'import setuptools, tokenize;file='"'"'C:\Users\John\AppData\Local\Temp\pip-install-4hgmcjg1\python-geohash\setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(file);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, file, '"'"'exec'"'"'))' bdist_wheel -d 'C:\Users\John\AppData\Local\Temp\pip-wheel-gijlszf9' --python-tag cp37:
ERROR: running bdist_wheel
running build
running build_py
creating build
creating build\lib.win-amd64-3.7
copying geohash.py -> build\lib.win-amd64-3.7
copying quadtree.py -> build\lib.win-amd64-3.7
copying jpgrid.py -> build\lib.win-amd64-3.7
copying jpiarea.py -> build\lib.win-amd64-3.7
running build_ext
building '_geohash' extension
error: Microsoft Visual C++ 14.0 is required. Get it with "Microsoft Visual C++ Build Tools": https://visualstudio.microsoft.com/downloads/

ERROR: Failed building wheel for python-geohash
Running setup.py clean for python-geohash
Failed to build python-geohash
ERROR: spyder 3.3.6 requires pyqt5<5.13; python_version >= "3", which is not installed.
ERROR: spyder 3.3.6 requires pyqtwebengine<5.13; python_version >= "3", which is not installed.
ERROR: distributed 2021.11.1 has requirement dask==2021.11.1, but you'll have dask 2.1.0 which is incompatible.
ERROR: woodwork 0.9.0 has requirement pandas>=1.3.0, but you'll have pandas 1.2.3 which is incompatible.
ERROR: featuretools 1.2.0 has requirement dask[dataframe]>=2021.10.0, but you'll have dask 2.1.0 which is incompatible.
ERROR: featuretools 1.2.0 has requirement pandas<2.0.0,>=1.3.0, but you'll have pandas 1.2.3 which is incompatible.
Installing collected packages: psutil, pyyaml, tblib, distributed, woodwork, featuretools, python-geohash, hypernets
Found existing installation: psutil 5.6.3
Uninstalling psutil-5.6.3:
Successfully uninstalled psutil-5.6.3
Found existing installation: PyYAML 5.1.1
ERROR: Cannot uninstall 'PyYAML'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall.

export the experiment report

The experiment report includes:

  • dataset information
  • feature transformation
  • evaluation indicators
  • confusion matrix
  • resources monitoring
  • importance of features
  • prediction speed
  • information of the ensemble models

The example of an exported experiment report:

  • example

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.