path stringlengths 13 17 | screenshot_names listlengths 1 873 | code stringlengths 0 40.4k | cell_type stringclasses 1
value |
|---|---|---|---|
17118075/cell_23 | [
"text_html_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.model_selection import cross_val_score
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
dataset = pd.get_dummies(df, columns=['sex', 'cp', 'fbs', 're... | code |
17118075/cell_33 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.linear_model import LogisticRegression
from sklearn.metrics import classification_report
lr = LogisticRegression()
lr.fit(X_train, y_train)
pred = lr.predict(X_test)
pred
from sklearn.metrics import classification_report
report = classification_repo... | code |
17118075/cell_20 | [
"text_plain_output_1.png"
] | from sklearn.model_selection import cross_val_score
from sklearn.tree import DecisionTreeClassifier
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
dataset = pd.get_dummies(df, columns=['sex', 'cp', 'fbs', 'restec... | code |
17118075/cell_6 | [
"text_html_output_1.png"
] | import pandas as pd
df = pd.read_csv('../input/heart.csv')
df.info() | code |
17118075/cell_29 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.linear_model import LogisticRegression
lr = LogisticRegression()
lr.fit(X_train, y_train)
pred = lr.predict(X_test)
pred | code |
17118075/cell_2 | [
"text_plain_output_1.png"
] | import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
import warnings
warnings.filterwarnings('ignore')
import os
print(os.listdir('../input')) | code |
17118075/cell_11 | [
"text_html_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
plt.figure(figsize=(10, 7))
sns.set_style('whitegrid')
sns.countplot('target', data=df) | code |
17118075/cell_19 | [
"text_html_output_1.png"
] | from sklearn.ensemble import RandomForestClassifier
from sklearn.model_selection import cross_val_score
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
dataset = pd.get_dummies(df, columns=['sex', 'cp', 'fbs', 're... | code |
17118075/cell_7 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import pandas as pd
df = pd.read_csv('../input/heart.csv')
df.describe() | code |
17118075/cell_32 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.linear_model import LogisticRegression
from sklearn.metrics import confusion_matrix
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
sns.set_style('... | code |
17118075/cell_28 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.linear_model import LogisticRegression
lr = LogisticRegression()
lr.fit(X_train, y_train) | code |
17118075/cell_8 | [
"text_plain_output_1.png"
] | import pandas as pd
df = pd.read_csv('../input/heart.csv')
df.shape | code |
17118075/cell_31 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.linear_model import LogisticRegression
from sklearn.metrics import confusion_matrix
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
dataset = pd.ge... | code |
17118075/cell_14 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
dataset = pd.get_dummies(df, columns=['sex', 'cp', 'fbs', 'restecg', 'exang', 'slope', 'ca', 'thal'])
dataset.head() | code |
17118075/cell_22 | [
"text_plain_output_1.png",
"image_output_1.png"
] | from sklearn.model_selection import cross_val_score
from sklearn.naive_bayes import GaussianNB
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
dataset = pd.get_dummies(df, columns=['sex', 'cp', 'fbs', 'restecg', '... | code |
17118075/cell_10 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
df = pd.read_csv('../input/heart.csv')
df.shape
df.target.value_counts()
plt.figure(figsize=(12, 7))
sns.heatmap(df.corr(), cmap='coolwarm', annot=True) | code |
2006619/cell_13 | [
"text_html_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
x=train.iloc[:,2:].sum()
#plot
plt.figure(figsize=(8,4))
ax= sns.barplot(x.i... | code |
2006619/cell_9 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
x=train.iloc[:,2:].sum()
#plot
plt.figure(figsize=(8,4))
ax= sns.barplot(x.i... | code |
2006619/cell_6 | [
"text_plain_output_1.png"
] | import pandas as pd
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
print(' : train : test')
print('rows :', nrow_train, ':', nrow_test)
print('perc :', round(nrow_train * 100 / sum), ' :',... | code |
2006619/cell_2 | [
"text_plain_output_1.png"
] | #Check the dataset sizes(in MB)
!du -l ../input/* | code |
2006619/cell_11 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
x=train.iloc[:,2:].sum()
#plot
plt.figure(figsize=(8,4))
ax= sns.barplot(x.i... | code |
2006619/cell_8 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
x = train.iloc[:, 2:].sum()
plt.figure(figsize=(8, 4))
ax = sns.barplot(x.in... | code |
2006619/cell_16 | [
"image_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
x=train.iloc[:,2:].sum()
#plot
plt.figure(figsize=(8,4))
ax= sns.barplot(x.i... | code |
2006619/cell_17 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
x=train.iloc[:,2:].sum()
#plot
plt.figure(figsize=(8,4))
ax= sns.barplot(x.i... | code |
2006619/cell_14 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
nrow_train = train.shape[0]
nrow_test = test.shape[0]
sum = nrow_train + nrow_test
x=train.iloc[:,2:].sum()
#plot
plt.figure(figsize=(8,4))
ax= sns.barplot(x.i... | code |
2006619/cell_5 | [
"text_plain_output_1.png"
] | import pandas as pd
train = pd.read_csv('../input/train.csv')
test = pd.read_csv('../input/test.csv')
train.tail(10) | code |
74067198/cell_9 | [
"text_plain_output_1.png"
] | from sklearn.feature_selection import SelectKBest, f_classif
from sklearn.impute import SimpleImputer
from sklearn.linear_model import LogisticRegression
from sklearn.model_selection import train_test_split
from sklearn.neural_network import MLPClassifier
from sklearn.pipeline import make_pipeline
from sklearn.pr... | code |
74067198/cell_4 | [
"text_plain_output_1.png"
] | import pandas as pd
train_df = pd.read_csv('../input/tabular-playground-series-sep-2021/train.csv')
test_df = pd.read_csv('../input/tabular-playground-series-sep-2021/test.csv')
sample_df = pd.read_csv('../input/tabular-playground-series-sep-2021/sample_solution.csv')
train_df.describe() | code |
74067198/cell_6 | [
"text_plain_output_1.png",
"image_output_1.png"
] | from sklearn.model_selection import train_test_split
import pandas as pd
train_df = pd.read_csv('../input/tabular-playground-series-sep-2021/train.csv')
test_df = pd.read_csv('../input/tabular-playground-series-sep-2021/test.csv')
sample_df = pd.read_csv('../input/tabular-playground-series-sep-2021/sample_solution.cs... | code |
74067198/cell_8 | [
"text_plain_output_1.png"
] | from sklearn.feature_selection import SelectKBest, f_classif
from sklearn.impute import SimpleImputer
from sklearn.linear_model import LogisticRegression
from sklearn.model_selection import train_test_split
from sklearn.pipeline import make_pipeline
from sklearn.preprocessing import StandardScaler
import pandas a... | code |
74067198/cell_10 | [
"text_html_output_1.png"
] | from sklearn.ensemble import GradientBoostingClassifier
from sklearn.feature_selection import SelectKBest, f_classif
from sklearn.impute import SimpleImputer
from sklearn.linear_model import LogisticRegression
from sklearn.model_selection import train_test_split
from sklearn.neural_network import MLPClassifier
fr... | code |
74067198/cell_5 | [
"text_plain_output_1.png"
] | import pandas as pd
train_df = pd.read_csv('../input/tabular-playground-series-sep-2021/train.csv')
test_df = pd.read_csv('../input/tabular-playground-series-sep-2021/test.csv')
sample_df = pd.read_csv('../input/tabular-playground-series-sep-2021/sample_solution.csv')
train_df.isna().sum().describe() | code |
90140085/cell_9 | [
"text_html_output_1.png"
] | from bs4 import BeautifulSoup
import pandas as pd
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/market/'
table0 = soup.find_all('a')
table1 = []
for item in table0:
if item.... | code |
90140085/cell_6 | [
"text_plain_output_1.png"
] | from bs4 import BeautifulSoup
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/market/'
table0 = soup.find_all('a')
table1 = []
for item in table0:
if item.text == '日足(CSV)':
... | code |
90140085/cell_2 | [
"text_plain_output_1.png"
] | !pip install requests
!pip install BeautifulSoup4 | code |
90140085/cell_11 | [
"text_plain_output_1.png"
] | from bs4 import BeautifulSoup
import pandas as pd
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/market/'
table0 = soup.find_all('a')
table1 = []
for item in table0:
if item.... | code |
90140085/cell_7 | [
"text_html_output_1.png"
] | from bs4 import BeautifulSoup
import pandas as pd
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/market/'
table0 = soup.find_all('a')
table1 = []
for item in table0:
if item.... | code |
90140085/cell_8 | [
"text_html_output_2.png"
] | from bs4 import BeautifulSoup
import pandas as pd
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/market/'
table0 = soup.find_all('a')
table1 = []
for item in table0:
if item.... | code |
90140085/cell_15 | [
"text_html_output_1.png"
] | from bs4 import BeautifulSoup
from plotly.subplots import make_subplots
import pandas as pd
import plotly.graph_objects as go
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/mar... | code |
90140085/cell_16 | [
"text_plain_output_1.png"
] | from bs4 import BeautifulSoup
from plotly.subplots import make_subplots
import pandas as pd
import plotly.graph_objects as go
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/mar... | code |
90140085/cell_10 | [
"text_plain_output_1.png"
] | from bs4 import BeautifulSoup
import pandas as pd
import requests
url = 'https://info.ctfx.jp/service/market/data_download.html'
res = requests.get(url)
soup = BeautifulSoup(res.text, 'html.parser')
DIR = 'https://info.ctfx.jp/service/market/'
table0 = soup.find_all('a')
table1 = []
for item in table0:
if item.... | code |
1009667/cell_2 | [
"text_html_output_1.png",
"text_plain_output_1.png"
] | from glob import glob
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.DataFrame([dict(path=c_path, image_name=os.path.basename(c_path), type_cat=os.path.basename(os.path.dirname(c_path))) for c_path in glob('../input/train/*/*')])
print('Total Samples', train_df.shape[0... | code |
1009667/cell_3 | [
"text_html_output_1.png"
] | from glob import glob
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.DataFrame([dict(path=c_path, image_name=os.path.basename(c_path), type_cat=os.path.basename(os.path.dirname(c_path))) for c_path in glob('../input/train/*/*')])
train_df.sample(3)
test_df = pd.DataFr... | code |
1009667/cell_5 | [
"text_html_output_1.png"
] | from glob import glob
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.DataFrame([dict(path=c_path, image_name=os.path.basename(c_path), type_cat=os.path.basename(os.path.dirname(c_path))) for c_path in glob('../input/train/*/*')])
train_df.sample(3)
test_df = pd.DataFr... | code |
73080780/cell_9 | [
"application_vnd.jupyter.stderr_output_9.png",
"application_vnd.jupyter.stderr_output_7.png",
"application_vnd.jupyter.stderr_output_11.png",
"text_plain_output_20.png",
"text_plain_output_4.png",
"text_plain_output_14.png",
"text_plain_output_10.png",
"text_plain_output_6.png",
"text_plain_output_2... | from sklearn import compose
from sklearn import impute
from sklearn import metrics
from sklearn import model_selection
from sklearn import pipeline
from sklearn import preprocessing
import lightgbm as lgbm
import numpy as np
import pandas as pd
import random
import xgboost as xgb
import glob
import pandas as... | code |
73080780/cell_11 | [
"text_plain_output_2.png",
"text_plain_output_1.png"
] | from sklearn import metrics
from sklearn import model_selection
import numpy as np
import pandas as pd
import random
FOLDS = 10
df_train = pd.read_csv('../input/30-days-of-ml/train.csv')
kfold = model_selection.KFold(n_splits=FOLDS, shuffle=True, random_state=42)
df_train['fold'] = -1
for fold, (_, valid_idx) in e... | code |
18159892/cell_21 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import cv2
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('../input/train.csv')
test_df = pd.read_csv('../input/test.csv')
def display_train_images(df, col=3, row=3):
tol = 30 # initializing... | code |
18159892/cell_4 | [
"application_vnd.jupyter.stderr_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('../input/train.csv')
test_df = pd.read_csv('../input/test.csv')
train_df.head() | code |
18159892/cell_34 | [
"text_plain_output_1.png"
] | from keras import layers
from keras.applications import DenseNet121
from keras.models import Sequential
from keras.optimizers import Adam
densenet = DenseNet121(weights='imagenet', include_top=False, input_shape=(224, 224, 3))
def build_model():
model = Sequential()
model.add(densenet)
model.add(layers... | code |
18159892/cell_1 | [
"application_vnd.jupyter.stderr_output_2.png",
"text_plain_output_1.png"
] | # This Python 3 environment comes with many helpful analytics libraries installed
# It is defined by the kaggle/python docker image: https://github.com/kaggle/docker-python
# For example, here's several helpful packages to load in
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import pandas as pd ... | code |
18159892/cell_16 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('../input/train.csv')
test_df = pd.read_csv('../input/test.csv')
train_df['diagnosis'].value_counts().plot.bar()
plt.title('Visualization of DR Trainig Dataset')
plt.xlabel('DR Diagnosis Type... | code |
18159892/cell_35 | [
"text_plain_output_1.png"
] | from keras import layers
from keras.applications import DenseNet121
from keras.callbacks import Callback, ModelCheckpoint
from keras.models import Sequential
from keras.optimizers import Adam
from keras.preprocessing.image import ImageDataGenerator
from sklearn.metrics import cohen_kappa_score, accuracy_score
BA... | code |
18159892/cell_31 | [
"image_output_1.png"
] | from keras.applications import DenseNet121
densenet = DenseNet121(weights='imagenet', include_top=False, input_shape=(224, 224, 3)) | code |
18159892/cell_24 | [
"image_output_1.png"
] | import cv2
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('../input/train.csv')
test_df = pd.read_csv('../input/test.csv')
def display_train_images(df, col=3, row=3):
tol = 30 # initializing... | code |
18159892/cell_14 | [
"text_html_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('../input/train.csv')
test_df = pd.read_csv('../input/test.csv')
train_df['diagnosis'].value_counts() | code |
32068582/cell_6 | [
"text_plain_output_1.png"
] | from scipy.optimize import curve_fit
from sklearn.metrics import mean_squared_error
import matplotlib.pyplot as plt
import numpy as np
import numpy as np # linear algebra
popDict = {'Afghanistan_NONE': 34656032, 'Albania_NONE': 2876101, 'Algeria_NONE': 40606052, 'Andorra_NONE': 77281, 'Angola_NONE': 28813463, 'Ant... | code |
32068582/cell_1 | [
"text_plain_output_1.png"
] | import os
import numpy as np
import pandas as pd
import os
for dirname, _, filenames in os.walk('/kaggle/input'):
for filename in filenames:
print(os.path.join(dirname, filename)) | code |
50221500/cell_9 | [
"text_html_output_1.png"
] | import pandas as pd
train_df = pd.read_csv('../input/titanic-machine-learning-from-disaster/train.csv')
test_df = pd.read_csv('../input/titanic-machine-learning-from-disaster/test.csv')
test_df.head() | code |
50221500/cell_1 | [
"text_plain_output_1.png"
] | import os
import os
for dirname, _, filenames in os.walk('/kaggle/input'):
for filename in filenames:
print(os.path.join(dirname, filename)) | code |
50221500/cell_8 | [
"text_html_output_1.png"
] | import pandas as pd
train_df = pd.read_csv('../input/titanic-machine-learning-from-disaster/train.csv')
test_df = pd.read_csv('../input/titanic-machine-learning-from-disaster/test.csv')
gender_df = pd.read_csv('../input/titanic-machine-learning-from-disaster/gender_submission.csv')
gender_df.head() | code |
50221500/cell_10 | [
"text_html_output_1.png"
] | import pandas as pd
train_df = pd.read_csv('../input/titanic-machine-learning-from-disaster/train.csv')
train_df.head() | code |
18143939/cell_13 | [
"application_vnd.jupyter.stderr_output_1.png"
] | import bs4
import pandas as pd
import requests
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/31/transcript-first-night-second-democratic-debate')
r2 = requests.get('https://www.washingtonpost.com/politics/2019/08/01/transcript-night-second-democratic-debate/')
def parse_requests(r, night=None):... | code |
18143939/cell_9 | [
"image_output_1.png"
] | import bs4
import requests
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/31/transcript-first-night-second-democratic-debate')
r2 = requests.get('https://www.washingtonpost.com/politics/2019/08/01/transcript-night-second-democratic-debate/')
def parse_requests(r, night=None):
soup = bs4.Beaut... | code |
18143939/cell_20 | [
"text_html_output_1.png"
] | from nltk.stem.porter import PorterStemmer
from wordcloud import WordCloud
import bs4
import matplotlib.pyplot as plt
import matplotlib.pyplot as plt
import nltk
import pandas as pd
import requests
import sklearn.feature_extraction.text as skt
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/... | code |
18143939/cell_6 | [
"text_plain_output_1.png"
] | import bs4
import pandas as pd
import requests
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/31/transcript-first-night-second-democratic-debate')
r2 = requests.get('https://www.washingtonpost.com/politics/2019/08/01/transcript-night-second-democratic-debate/')
def parse_requests(r, night=None):... | code |
18143939/cell_7 | [
"image_output_1.png"
] | import bs4
import pandas as pd
import requests
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/31/transcript-first-night-second-democratic-debate')
r2 = requests.get('https://www.washingtonpost.com/politics/2019/08/01/transcript-night-second-democratic-debate/')
def parse_requests(r, night=None):... | code |
18143939/cell_17 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from nltk.stem.porter import PorterStemmer
from wordcloud import WordCloud
import bs4
import matplotlib.pyplot as plt
import matplotlib.pyplot as plt
import nltk
import pandas as pd
import requests
import sklearn.feature_extraction.text as skt
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/... | code |
18143939/cell_10 | [
"text_plain_output_1.png"
] | import bs4
import pandas as pd
import requests
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/31/transcript-first-night-second-democratic-debate')
r2 = requests.get('https://www.washingtonpost.com/politics/2019/08/01/transcript-night-second-democratic-debate/')
def parse_requests(r, night=None):... | code |
18143939/cell_12 | [
"text_plain_output_1.png"
] | import bs4
import pandas as pd
import requests
r1 = requests.get('https://www.washingtonpost.com/politics/2019/07/31/transcript-first-night-second-democratic-debate')
r2 = requests.get('https://www.washingtonpost.com/politics/2019/08/01/transcript-night-second-democratic-debate/')
def parse_requests(r, night=None):... | code |
105198632/cell_4 | [
"text_plain_output_1.png"
] | j = 0
for i in range(1, 31):
j = j + i
i = 0
j = 0
while i <= 30:
j = j + i
i = i + 1
print(j) | code |
105198632/cell_2 | [
"text_plain_output_1.png"
] | j = 0
for i in range(1, 31):
j = j + i
print(j) | code |
309674/cell_3 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import networkx as nx
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import networkx as nx
import matplotlib.pyplot as plt
from subprocess import check_output
comments = pd.read_csv('../input/comment.csv')
likes = pd.read_... | code |
74045780/cell_9 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.dtypes
df.shape
df.AMR.value_counts(ascending=True)
df.CRISPR_Cas.value_counts(ascending=True) | code |
74045780/cell_4 | [
"text_plain_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.dtypes | code |
74045780/cell_6 | [
"text_plain_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.dtypes
df.shape
df.AMR.plot() | code |
74045780/cell_2 | [
"text_html_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.head() | code |
74045780/cell_1 | [
"text_plain_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
for dirname, _, filenames in os.walk('/kaggle/input'):
for filename in filenames:
print(os.path.join(dirname, filename))
df = pd.read_csv('../input/amr... | code |
74045780/cell_7 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.dtypes
df.shape
df.AMR.value_counts(ascending=True) | code |
74045780/cell_8 | [
"text_plain_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.dtypes
df.shape
df.AMR.value_counts(ascending=True)
df.CRISPR_Cas.plot() | code |
74045780/cell_3 | [
"text_plain_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.describe() | code |
74045780/cell_10 | [
"text_html_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.dtypes
df.shape
df.AMR.value_counts(ascending=True)
df.CRISPR_Cas.value_counts(ascending=True)
df.... | code |
74045780/cell_5 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
import matplotlib as plt
import os
df = pd.read_csv('../input/amrc-data/Efaecium_AMRC.csv')
df.dtypes
df.shape | code |
128044853/cell_2 | [
"text_html_output_1.png",
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.preprocessing import LabelEncoder
import nltk
import os
import numpy as np
import pandas as pd
from sklearn.preprocessing import LabelEncoder
from sklearn.model_selection import GridSearchCV
from sklearn.model_selection import train_test_split
import lightgbm as lgb
import nltk
nltk.download('punkt')
nl... | code |
128044853/cell_7 | [
"text_html_output_1.png"
] | from sklearn.model_selection import KFold
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import LabelEncoder
from tqdm import tqdm
import joblib
import lightgbm as lgb
import nltk
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O ... | code |
128044853/cell_8 | [
"text_plain_output_1.png"
] | from sklearn.model_selection import KFold
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import LabelEncoder
from tqdm import tqdm
import lightgbm as lgb
import nltk
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_c... | code |
128044853/cell_10 | [
"application_vnd.jupyter.stderr_output_2.png",
"text_plain_output_1.png"
] | from nltk.corpus import stopwords
from nltk.sentiment import SentimentIntensityAnalyzer
from nltk.tokenize import sent_tokenize
from nltk.tokenize import word_tokenize
from sklearn.model_selection import KFold
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import LabelEncoder
fro... | code |
128044853/cell_5 | [
"text_plain_output_1.png"
] | from sklearn.preprocessing import LabelEncoder
from tqdm import tqdm
import nltk
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import numpy as np
import pandas as pd
from sklearn.preprocessing import LabelEncoder
from sklearn.model_selection import GridSearchCV
from sklearn.model... | code |
122256136/cell_13 | [
"image_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_6 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
sns.countp... | code |
122256136/cell_2 | [
"text_plain_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
data.head() | code |
122256136/cell_1 | [
"text_plain_output_1.png"
] | import os
import numpy as np
import pandas as pd
import os
for dirname, _, filenames in os.walk('/kaggle/input'):
for filename in filenames:
print(os.path.join(dirname, filename)) | code |
122256136/cell_18 | [
"image_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_8 | [
"text_html_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_15 | [
"image_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_16 | [
"image_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_17 | [
"image_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_14 | [
"text_html_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_10 | [
"text_html_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
122256136/cell_12 | [
"image_output_1.png"
] | import matplotlib.pyplot as plt
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
data = pd.read_csv('/kaggle/input/e-commerce-shoppers-behaviour-understanding/train_data_v2.csv')
pd.set_option('display.max_columns', None)
plt.rcParams['figure.figsize'] = (8, 6)
n_missing... | code |
2010673/cell_13 | [
"text_html_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import nltk
import os
import pandas as pd
path = '../input/state-of-the-union-corpus-1989-2017'
dirs = os.listdir(path)
def count_words(word, filename):
file = open(path + '/' + filename, encoding='utf8')
text = file.read().lower()
words ... | code |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.