Skip to main content

Kaggle dataset download in colab

 import os

os.environ['KAGGLE_CONFIG_DIR']='/content/drive/MyDrive/Kaggle_Data'


%cd /content/drive/MyDrive/Kaggle_Data

/content/drive/MyDrive/Kaggle_Data


!ls

kaggle.json

!kaggle competitions download -c 30-days-of-ml # copy api from kaggel

2s
Warning: Looks like you're using an outdated API Version, please consider updating (server 1.5.12 / client 1.5.4)
Downloading sample_submission.csv.zip to /content/drive/My Drive/Kaggle_Data
  0% 0.00/470k [00:00<?, ?B/s]
100% 470k/470k [00:00<00:00, 32.6MB/s]
Downloading test.csv.zip to /content/drive/My Drive/Kaggle_Data
 87% 22.0M/25.2M [00:00<00:00, 104MB/s] 
100% 25.2M/25.2M [00:00<00:00, 98.3MB/s]
Downloading train.csv.zip to /content/drive/My Drive/Kaggle_Data
 77% 31.0M/40.3M [00:00<00:00, 103MB/s]
100% 40.3M/40.3M [00:00<00:00, 100MB/s]


!mv train.csv.zip 30-days-of-ml # moving file into folder
0s

[14]
0s
/content/drive/My Drive/Kaggle_Data/30-days-of-ml

[17]
10s
Archive:  train.csv.zip
replace train.csv? [y]es, [n]o, [A]ll, [N]one, [r]ename: n

[16]
1s
Archive:  test.csv.zip
  inflating: test.csv   







Popular posts from this blog

Bagging and Boosting

  What is an Ensemble Method? The ensemble is a method used in the machine learning algorithm. In this method, multiple models or ‘weak learners’ are trained to rectify the same problem and integrated to gain desired results. Weak models combined rightly give accurate models. Bagging Bagging is an acronym for ‘Bootstrap Aggregation’ and is used to decrease the variance in the prediction model. Bagging is a parallel method that fits different, considered learners independently from each other, making it possible to train them simultaneously. Bagging generates additional data for training from the dataset. This is achieved by random sampling with replacement from the original dataset. Sampling with replacement may repeat some observations in each new training data set. Every element in Bagging is equally probable for appearing in a new dataset.  These multi datasets are used to train multiple models in parallel. The average of all the predictions from different ensemble models i...