Github Stars Predictor
We have made a github repo star predictor that tries to predict the stars of any github repository having greater than 100 stars. It predicts based on the owner/organization’s status and activities (commits, forks, comments, branches, update rate, etc.) on the repository. Different types of models (Gradient boost, Deep neural network, etc) have been tested successfully on the dataset we fetched from github apis.
Project link: https://github.com/Doodies/Github-Stars-Predictor
- Pawan Bansal (https://github.com/hackbansu)
- Anmol Dua (https://github.com/anmolduainter)
- Himanshu Arora (https://github.com/himanshuarora97)
We used the github REST api and GraphQL api to collect data of repositories having more than 100 stars. The data is available in the dataset directory We were able to collect the data faster using the Digital Ocean’s multiple servers. So we thanks Digital Ocean for providing free credits to students to use servers. For the details on dataset features refer the summary section below.
- Python 2.7
- Jupyter Notebook
- Cat Boost
- Matplot Lib
We also used Google Colab’s GPU notebooks. So we thank to Google for starting thier colab project for providing GPUs
Below is a brief description for the Code files/folder in repo.
This is used for configuring the digital ocean server
This script fetches the basic info of repos having more than 100 stars using the Github REST API
This script fetches the complete info of the repositories that were fetched by the above script and uses the Github GraphQL API. It follows the approach of fetching the data at the fixed rate defined in env file (eg. 730ms per request)
This script fetches the complete info of the repositories that were fetched by the above script and uses the Github GraphQL API. It follows the approach of requesting data for next repo after receiving the response of the already sent request.
This script converts the json data fetched from Github’s GraphQL API in the above script to the equivalent csv file.
This scripts merges all the data in multiple csv files to a single csv file
We have done the feature engineering task in this notebook. It visualises the data and correspondingly creates new features, modifies existing features and removes redundant features. For details on features created, check the summary below
In this notebook, we trained different models with hyper parameter tuning on our dataset and compared their result in the end. For details on models trained, their prediction scores, etc. check the summary below.
In this project we have tried to predict the number of stars of a github repository that have more than 100 stars. For this we have taken the github repository data from github REST api and GraphQL api. After generating the dataset we visualized and did some feature engineering with the dataset and after that , finally we come up to the stage where we applied various models and predicted the model’s scores on training and test data.
There are total of 49 features before pre-processing. After pre-processing (adding new features, removal of redundant features and modifying existing features) the count changes to 54. All the features are listed below. Some features after pre-processing may not be clear. Please refer to the VisualizePreprocess.ipynb notebook for details.
Features after pre-processing
- Gradient Boost Regressor
- Cat Boost Regressor
- Random Forest Regressor
- Deep Neural Network
- R² score