For more details, please refer to Features.Experiments on public datasets show that LightGBM can outperform existing boosting frameworks on both efficiency and accuracy, with significantly lower memory consumption. What's more, the experiments show that LightGBM can achieve a linear speed-up by using multiple machines for training in specific settings.
gbdt gbm machine-learning data-mining kaggle efficiency distributed lightgbm gbrtIPython Notebook(s) demonstrating deep learning functionality.IPython Notebook(s) demonstrating scikit-learn functionality.
machine-learning deep-learning data-science big-data aws tensorflow theano caffe scikit-learn kaggle spark mapreduce hadoop matplotlib pandas numpy scipy kerasCatBoost is a machine learning method based on gradient boosting over decision trees. All CatBoost documentation is available here.
machine-learning decision-trees gradient-boosting gbm gbdt r kaggle gpu-computing catboost tutorial categorical-features distributed gpu coreml opensource data-science big-dataTel-Aviv Deep Learning Bootcamp is an intensive (and free!) 5-day program intended to teach you all about deep learning. It is nonprofit focused on advancing data science education and fostering entrepreneurship. The Bootcamp is a prominent venue for graduate students, researchers, and data science professionals. It offers a chance to study the essential and innovative aspects of deep learning. Participation is via a donation to the A.L.S ASSOCIATION for promoting research of the Amyotrophic Lateral Sclerosis (ALS) disease.
gpu nvidia docker-image machine-learning deep-learning data-science cuda-kernels kaggle-competition cuda pytorch pytorch-tutorials pytorch-tutorial bootcamp meetup kaggle kaggle-scripts pycudaA comprehensive list of Deep Learning / Artificial Intelligence and Machine Learning tutorials - rapidly expanding into areas of AI/Deep Learning / Machine Vision / NLP and industry specific areas such as Climate / Energy, Automotives, Retail, Pharma, Medicine, Healthcare, Policy, Ethics and more.
machine-learning deep-learning tensorflow pytorch keras matplotlib aws kaggle pandas scikit-learn torch artificial-intelligence neural-network convolutional-neural-networks tensorflow-tutorials python-data ipython-notebook capsule-networkI just built out v2 of this project that now gives you analytics info from your models, and is production-ready. machineJS is an amazing research project that clearly proved there's a hunger for automated machine learning. auto_ml tackles this exact same goal, but with more features, cleaner code, and the ability to be copy/pasted into production.
machine-learning data-science machine-learning-library machine-learning-algorithms ml data-scientists javascript-library scikit-learn kaggle numerai automated-machine-learning automl auto-ml neuralnet neural-network algorithms random-forest svm naive-bayes bagging optimization brainjs date-night sklearn ensemble data-formatting js xgboost scikit-neuralnetwork knn k-nearest-neighbors gridsearch gridsearchcv grid-search randomizedsearchcv preprocessing data-formatter kaggle-competitionPlease check your Torch7/CUDA environment when this code fails. Place the data files into a subfolder ./data.
kaggleAmazon Forest Computer Vision: Satellite Image tagging code using PyTorch / Keras with lots of PyTorch tricks
pytorch data-augmentation kaggle-competition kaggle deep-learning computer-vision keras neural-networks neural-network-example transfer-learningThis is an open solution to the Home Credit Default Risk challenge 🏡. In this open source solution you will find references to the neptune.ml. It is free platform for community Users, which we use daily to keep track of our experiments. Please note that using neptune.ml is not necessary to proceed with this solution. You may run it as plain Python script 🐍.
machine-learning deep-learning kaggle pipeline feature-engineering reproducible-experiments reproducibility pipeline-framework lightgbm xgboost neptune competition credit-scoring credit-risk open-source python3 python35This repository contains a 1st place solution for the Painter by Numbers competition on Kaggle. Below is a brief description of the dataset and approaches I've used to build and validate a predictive model. The challenge of the competition was to examine pairs of paintings and determine whether they were painted by the same artist. The training set consists of artwork images and their corresponding class labels (painters). Examples in the test set were split into 13 groups and all possible pairs within each group needed to be examined for the submission. The evaluation metric for the leaderboard was AUC (area under the curve).
machine-learning deep-learning kaggle neural-networks competitionIn this repo we compare two of the fastest boosted decision tree libraries: XGBoost and LightGBM. We will evaluate them across datasets of several domains and different sizes.On July 25, 2017, we published a blog post evaluating both libraries and discussing the benchmark results. The post is Lessons Learned From Benchmarking Fast Machine Learning Algorithms.
lightgbm xgboost boosted-trees machine-learning gpu benchmark azure distributed-systems gbdt gbm gbrt kaggleThis project and the data explores the relationship between Social Media, Salary, Influence, Performance and Team Valuation in the NBA.
r nba ggplot2 ipython-notebook jupyter-notebook machine-learning machine-learning-algorithms ml prediction social-network social-media salary influence court-performance pie social kaggle kaggle-dataset social-networksConvolutional neural networks for Google speech commands data set with PyTorch. We, xuyuan and tugstugi, have participated in the Kaggle competition TensorFlow Speech Recognition Challenge and reached the 10-th place. This repository contains a simplified and cleaned up version of our team's code.
speech-recognition deep-learning cifar10 neural-network kaggle classification pytorch resnet resnext densenet wide-residual-networks dual-path-networks2nd place solution for Airbnb New User Bookings Competition. Note: This code should be differ from my submitted solution(Public:0.88209/Private:0.88682) because of the seed settings. if you select a model of more than 5 fold-CV 0.833600, you can get about 0.88682(Private).
kaggle kaggle-competitionAll my submissions for Kaggle contests that I have been, and going to be participating. I will probably have everything written in Python (utilizing scikit-learn or similar libraries), but occasionally I might also use R or Haskell if I can.
kaggle xgboost mnist neural-network scikit-learn kaggle-contestMy goal for this minimal data science blog series is not only sharing, tutorializing, but also, making personal notes while learning and working as a Data Scientist. I’m looking forward to receiving any feedback from you. Chapter-1: Classify StarCraft 2 players with Python Pandas and Scikit-learn.
blog-series data-science scikit-learn xgboost kaggle machine-learningKaggle "Microsoft Malware Classification Challenge". 6th place solution
kaggle malware machine-learningCode for Coupon Purchase Prediction (RECRUIT Challenge). Note: This code is able to achieve a 5th place score (Private LB: 0.008776). But this is not a full version of my submitted solution (Private LB: 0.008905). My submitted solution is average of this solution and another XGBoost solution. This repositoy provides a simple version of 5th place solution.
kaggle
We have large collection of open source products. Follow the tags from
Tag Cloud >>
Open source products are scattered around the web. Please provide information
about the open source projects you own / you use.
Add Projects.