path stringlengths 13 17 | screenshot_names listlengths 1 873 | code stringlengths 0 40.4k | cell_type stringclasses 1
value |
|---|---|---|---|
33096987/cell_20 | [
"text_plain_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
train_df[['Pclass', 'Survived']].groupby(['Pclass'], as_index=False).mean... | code |
33096987/cell_6 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
train_df.head() | code |
33096987/cell_39 | [
"text_html_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
plt.style.use('seaborn-w... | code |
33096987/cell_26 | [
"text_html_output_1.png"
] | from collections import Counter
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
def... | code |
33096987/cell_48 | [
"text_html_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
p... | code |
33096987/cell_2 | [
"text_html_output_1.png"
] | import matplotlib.pyplot as plt
import os
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
plt.style.use('seaborn-whitegrid')
from collections import Counter
import warnings
warnings.filterwarnings('ignore')
import os
for dirname, _, filenames in os.walk('/... | code |
33096987/cell_7 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
train_df.describe() | code |
33096987/cell_45 | [
"text_html_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
p... | code |
33096987/cell_18 | [
"text_html_output_1.png"
] | import matplotlib.pyplot as plt
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
plt.style.use('seaborn-whitegrid')
from collections import Counter
import warnings
warnings.fi... | code |
33096987/cell_32 | [
"text_html_output_1.png"
] | from collections import Counter
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
def... | code |
33096987/cell_51 | [
"text_plain_output_1.png",
"image_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
p... | code |
33096987/cell_15 | [
"text_plain_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
category2 = ['Cabin', 'Name', 'Ticket']
for i in category2:
print(f'{... | code |
33096987/cell_35 | [
"text_html_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
plt.style.use('seaborn-w... | code |
33096987/cell_31 | [
"text_html_output_1.png"
] | from collections import Counter
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
def... | code |
33096987/cell_14 | [
"text_plain_output_1.png"
] | import matplotlib.pyplot as plt
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
plt.style.use('seaborn-whitegrid')
from collections import Counter
import warnings
warnings.fi... | code |
33096987/cell_22 | [
"text_plain_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns
train_df[['SibSp', 'Survived']].groupby(['SibSp'], as_index=False).mean()... | code |
33096987/cell_37 | [
"text_plain_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
plt.style.use('seaborn-w... | code |
33096987/cell_5 | [
"text_html_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
train_df = pd.read_csv('/kaggle/input/titanic/train.csv')
test_df = pd.read_csv('/kaggle/input/titanic/test.csv')
test_PassengerId = test_df['PassengerId']
train_df.columns | code |
33096987/cell_36 | [
"text_plain_output_1.png"
] | from collections import Counter
import matplotlib.pyplot as plt
import numpy as np # linear algebra
import os
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import warnings
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
plt.style.use('seaborn-w... | code |
90148984/cell_13 | [
"text_plain_output_1.png",
"image_output_1.png"
] | dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
len(y) | code |
90148984/cell_9 | [
"text_plain_output_1.png",
"image_output_1.png"
] | dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
len(df) | code |
90148984/cell_25 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LinearRegression
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
model = LinearRegression()
model.fit(X, y)
model.score(X, y)
model.intercept_
model.coef_
X_test = df.drop(columns=['... | code |
90148984/cell_23 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LinearRegression
import pandas as pd
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
model = LinearRegression()
model.fit(X, y)
model.score(X, y)
model.intercept_
model.coef_
X_test... | code |
90148984/cell_30 | [
"text_html_output_1.png"
] | import numpy as np
import seaborn as sns
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
with sns.plotting_context("notebook",font_scale=2.5):
g = sns.pairplot(dataset[['sqft_lot','sqft_above','price','sqft_living','bedrooms']],
hue='bedrooms', palette='tab20',heigh... | code |
90148984/cell_20 | [
"text_plain_output_1.png"
] | dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X_test = df.drop(columns=['price'])[:10]
X_test | code |
90148984/cell_6 | [
"text_plain_output_1.png"
] | import seaborn as sns
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
sns.lineplot(x='yr_built', y='sqft_living', data=df, ci=None) | code |
90148984/cell_29 | [
"text_html_output_1.png"
] | import numpy as np
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X_test = df.drop(columns=['price'])[:10]
X_test
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X = df.drop(columns=['price'])[:10]
y = df['p... | code |
90148984/cell_2 | [
"text_plain_output_1.png"
] | dataset.columns | code |
90148984/cell_19 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LinearRegression
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
model = LinearRegression()
model.fit(X, y)
model.score(X, y)
model.intercept_
model.coef_ | code |
90148984/cell_1 | [
"text_html_output_1.png"
] | import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
from sklearn.linear_model import LinearRegression
dataset = pd.read_csv('../input/kc-house-data/kc_house_data.csv')
dataset.head() | code |
90148984/cell_7 | [
"text_plain_output_1.png"
] | import seaborn as sns
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
sns.lmplot(x='bedrooms', y='price', data=df, ci=None) | code |
90148984/cell_18 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LinearRegression
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
model = LinearRegression()
model.fit(X, y)
model.score(X, y)
model.intercept_ | code |
90148984/cell_28 | [
"text_plain_output_1.png"
] | import numpy as np
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X_test = df.drop(columns=['price'])[:10]
X_test
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X = df.drop(columns=['price'])[:10]
y = df['p... | code |
90148984/cell_8 | [
"text_plain_output_1.png"
] | import seaborn as sns
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
with sns.plotting_context('notebook', font_scale=2.5):
g = sns.pairplot(dataset[['sqft_lot', 'sqft_above', 'price', 'sqft_living', 'bedrooms']], hue='bedrooms', palette='tab20', height=6)
g.set(xticklabels=[]) | code |
90148984/cell_15 | [
"text_plain_output_1.png",
"image_output_1.png"
] | dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X.head() | code |
90148984/cell_16 | [
"image_output_1.png"
] | from sklearn.linear_model import LinearRegression
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
model = LinearRegression()
model.fit(X, y) | code |
90148984/cell_3 | [
"text_plain_output_1.png"
] | dataset.columns
print(dataset.dtypes) | code |
90148984/cell_17 | [
"text_plain_output_1.png"
] | from sklearn.linear_model import LinearRegression
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
model = LinearRegression()
model.fit(X, y)
model.score(X, y) | code |
90148984/cell_14 | [
"text_plain_output_1.png",
"image_output_1.png"
] | dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
y.head() | code |
90148984/cell_22 | [
"text_html_output_1.png"
] | from sklearn.linear_model import LinearRegression
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
model = LinearRegression()
model.fit(X, y)
model.score(X, y)
model.intercept_
model.coef_
X_test = df.drop(columns=['... | code |
90148984/cell_10 | [
"text_plain_output_1.png"
] | dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns | code |
90148984/cell_27 | [
"text_plain_output_1.png"
] | import numpy as np
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X_test = df.drop(columns=['price'])[:10]
X_test
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
X = df.drop(columns=['price'])[:10]
y = df['p... | code |
90148984/cell_12 | [
"text_plain_output_1.png"
] | dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
df.columns
X = df.drop(columns=['price'])[:10]
y = df['price'][:10]
len(X) | code |
90148984/cell_5 | [
"text_plain_output_1.png"
] | import seaborn as sns
dataset.columns
df = dataset.drop(columns=['date', 'id'])
df1 = df.dropna()
sns.lmplot(x='price', y='sqft_living', data=df, ci=None) | code |
50212838/cell_13 | [
"text_html_output_1.png"
] | from sklearn.cluster import MiniBatchKMeans
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.metrics.pairwise import linear_kernel
import networkx as nx
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv'... | code |
50212838/cell_9 | [
"text_plain_output_1.png"
] | from sklearn.cluster import MiniBatchKMeans
from sklearn.feature_extraction.text import TfidfVectorizer
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv')
data['date_added'] = pd.to_datetime(data['date_added'])
data['year'] = d... | code |
50212838/cell_1 | [
"text_plain_output_1.png"
] | import os
import numpy as np
import pandas as pd
import os
for dirname, _, filenames in os.walk('/kaggle/input'):
for filename in filenames:
print(os.path.join(dirname, filename)) | code |
50212838/cell_7 | [
"text_plain_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv')
data['date_added'] = pd.to_datetime(data['date_added'])
data['year'] = data['date_added'].dt.year
data['month'] = data['date_added'].dt.month
data['day'] = data['date_added'].dt.... | code |
50212838/cell_8 | [
"image_output_1.png"
] | from sklearn.cluster import MiniBatchKMeans
from sklearn.feature_extraction.text import TfidfVectorizer
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv')
data['date_added'] = pd.to_datetime(data['date_added'])
data['year'] = d... | code |
50212838/cell_16 | [
"text_plain_output_1.png"
] | from sklearn.cluster import MiniBatchKMeans
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.metrics.pairwise import linear_kernel
import matplotlib.pyplot as plt
import networkx as nx
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/... | code |
50212838/cell_3 | [
"text_plain_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv')
data.head() | code |
50212838/cell_14 | [
"text_html_output_1.png"
] | from sklearn.cluster import MiniBatchKMeans
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.metrics.pairwise import linear_kernel
import networkx as nx
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv'... | code |
50212838/cell_10 | [
"text_html_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv')
data['date_added'] = pd.to_datetime(data['date_added'])
data['year'] = data['date_added'].dt.year
data['month'] = data['date_added'].dt.month
data['day'] = data['date_added'].dt.... | code |
50212838/cell_5 | [
"text_plain_output_1.png"
] | import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
data = pd.read_csv('/kaggle/input/netflix-shows/netflix_titles.csv')
data.describe() | code |
1008986/cell_9 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.metrics import confusion_matrix, classification_report, roc_curve, roc_auc_score
from sklearn.model_selection import StratifiedKFold
from sklearn.preprocessing import StandardScaler
import itertools
import matplotlib.pyplot as plt
import numpy as np... | code |
1008986/cell_4 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.preprocessing import StandardScaler
import itertools
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
from sklearn.linear_model import LogisticRegression
from sklearn.preprocessing import StandardScaler
from s... | code |
1008986/cell_20 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.metrics import confusion_matrix, classification_report, roc_curve, roc_auc_score
from sklearn.model_selection import StratifiedKFold
from sklearn.preprocessing import StandardScaler
import itertools
import matplotlib.pyplot as plt
import numpy as np... | code |
1008986/cell_6 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.preprocessing import StandardScaler
import itertools
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
from sklearn.linear_model import LogisticRegression
from sklearn.preprocessing import StandardScaler
from s... | code |
1008986/cell_18 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.metrics import confusion_matrix, classification_report, roc_curve, roc_auc_score
from sklearn.model_selection import StratifiedKFold
from sklearn.preprocessing import StandardScaler
import itertools
import matplotlib.pyplot as plt
import numpy as np... | code |
1008986/cell_16 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.metrics import confusion_matrix, classification_report, roc_curve, roc_auc_score
from sklearn.model_selection import StratifiedKFold
from sklearn.preprocessing import StandardScaler
import itertools
import matplotlib.pyplot as plt
import numpy as np... | code |
1008986/cell_22 | [
"application_vnd.jupyter.stderr_output_1.png"
] | from sklearn.linear_model import LogisticRegression
from sklearn.metrics import confusion_matrix, classification_report, roc_curve, roc_auc_score
from sklearn.model_selection import StratifiedKFold
from sklearn.preprocessing import StandardScaler
import itertools
import matplotlib.pyplot as plt
import numpy as np... | code |
90112109/cell_13 | [
"text_html_output_1.png"
] | import pandas as pd
import seaborn as sb
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique()
data.isnull().sum()
data.corr()
cor = data.corr()
data2 = data.drop(['children', 'region'], axis=1)
sb.relplot(x='age', y='charges', hue='smoker', data=data2) | code |
90112109/cell_9 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import pandas as pd
import seaborn as sb
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique()
data.isnull().sum()
data.corr()
cor = data.corr()
sb.heatmap(cor, xticklabels=cor.columns, yticklabels=cor.columns, annot=True) | code |
90112109/cell_4 | [
"text_html_output_1.png"
] | import pandas as pd
data = pd.read_csv('../input/insurance/insurance.csv')
data.info() | code |
90112109/cell_6 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import pandas as pd
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique() | code |
90112109/cell_7 | [
"text_html_output_1.png"
] | import pandas as pd
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique()
data.isnull().sum() | code |
90112109/cell_8 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import pandas as pd
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique()
data.isnull().sum()
data.corr() | code |
90112109/cell_3 | [
"text_plain_output_1.png"
] | import pandas as pd
data = pd.read_csv('../input/insurance/insurance.csv')
data.head() | code |
90112109/cell_14 | [
"text_plain_output_1.png"
] | import pandas as pd
import seaborn as sb
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique()
data.isnull().sum()
data.corr()
cor = data.corr()
data2 = data.drop(['children', 'region'], axis=1)
sb.displot(data2['charges']) | code |
90112109/cell_10 | [
"text_html_output_1.png"
] | import pandas as pd
import seaborn as sb
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique()
data.isnull().sum()
data.corr()
cor = data.corr()
sb.pairplot(data) | code |
90112109/cell_12 | [
"text_plain_output_1.png"
] | import pandas as pd
import seaborn as sb
data = pd.read_csv('../input/insurance/insurance.csv')
data.nunique()
data.isnull().sum()
data.corr()
cor = data.corr()
data2 = data.drop(['children', 'region'], axis=1)
data2.head() | code |
90112109/cell_5 | [
"text_plain_output_1.png",
"image_output_1.png"
] | import pandas as pd
data = pd.read_csv('../input/insurance/insurance.csv')
data.describe() | code |
73074228/cell_13 | [
"text_html_output_1.png"
] | from pathlib import Path
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
train.shape
train.info() | code |
73074228/cell_20 | [
"text_plain_output_1.png"
] | from pathlib import Path
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
train.shape
train.isnull().any().sum()
y = train['loss']
y.head() | code |
73074228/cell_55 | [
"text_html_output_1.png"
] | from pathlib import Path
from sklearn.metrics import mean_squared_error
from sklearn.model_selection import train_test_split, GridSearchCV, RandomizedSearchCV, KFold
from sklearn.preprocessing import StandardScaler
from xgboost import XGBRegressor
import numpy as np
import pandas as pd
INPUT = Path('../input/tab... | code |
73074228/cell_54 | [
"text_plain_output_1.png"
] | from pathlib import Path
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
submission = pd.read_csv(INPUT / 'sample_submission.csv')
submission.head() | code |
73074228/cell_11 | [
"text_plain_output_1.png"
] | from pathlib import Path
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
train.shape
train.head() | code |
73074228/cell_32 | [
"text_plain_output_1.png"
] | from pathlib import Path
from sklearn.model_selection import train_test_split, GridSearchCV, RandomizedSearchCV, KFold
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
train.shape
train.isnull().any().sum... | code |
73074228/cell_8 | [
"text_html_output_1.png"
] | from xgboost import XGBRegressor
from catboost import CatBoostRegressor
from lightgbm import LGBMRegressor | code |
73074228/cell_15 | [
"text_plain_output_1.png"
] | from pathlib import Path
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
test.isnull().any().sum() | code |
73074228/cell_46 | [
"text_html_output_1.png"
] | from pathlib import Path
from sklearn.metrics import mean_squared_error
from sklearn.model_selection import train_test_split, GridSearchCV, RandomizedSearchCV, KFold
from sklearn.preprocessing import StandardScaler
from xgboost import XGBRegressor
import pandas as pd
INPUT = Path('../input/tabular-playground-seri... | code |
73074228/cell_14 | [
"text_plain_output_1.png"
] | from pathlib import Path
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
train.shape
train.isnull().any().sum() | code |
73074228/cell_12 | [
"text_html_output_1.png"
] | from pathlib import Path
import pandas as pd
INPUT = Path('../input/tabular-playground-series-aug-2021')
train = pd.read_csv(INPUT / 'train.csv')
test = pd.read_csv(INPUT / 'test.csv')
test.head() | code |
17137542/cell_21 | [
"image_output_1.png"
] | from sklearn import manifold
from sklearn.cluster import KMeans
from sklearn.tree import DecisionTreeClassifier
import matplotlib.pyplot as plt
import pandas as pd
df = pd.read_csv('../input/whisky.csv')
dist = []
for i in range(2, 20):
km = KMeans(n_clusters=i, n_init=10, max_iter=500, random_state=0)
km... | code |
17137542/cell_4 | [
"text_plain_output_1.png"
] | import pandas as pd
df = pd.read_csv('../input/whisky.csv')
df.head() | code |
17137542/cell_23 | [
"text_html_output_1.png"
] | from IPython.display import Image, display_png
from pydotplus import graph_from_dot_data
from sklearn import manifold
from sklearn.cluster import KMeans
from sklearn.tree import DecisionTreeClassifier
from sklearn.tree import export_graphviz
import matplotlib.pyplot as plt
import pandas as pd
df = pd.read_csv('... | code |
17137542/cell_6 | [
"image_output_1.png"
] | import pandas as pd
df = pd.read_csv('../input/whisky.csv')
df.info() | code |
17137542/cell_26 | [
"text_html_output_1.png"
] | from IPython.display import Image, display_png
from pydotplus import graph_from_dot_data
from pyproj import Proj, transform
from sklearn import manifold
from sklearn.cluster import KMeans
from sklearn.tree import DecisionTreeClassifier
from sklearn.tree import export_graphviz
import folium
import matplotlib.pyp... | code |
17137542/cell_2 | [
"text_plain_output_1.png"
] | !pip install pydotplus
import pandas as pd
import numpy as np
from sklearn.cluster import KMeans
import matplotlib.pyplot as plt
from sklearn.tree import DecisionTreeClassifier
from IPython.display import Image, display_png
from pydotplus import graph_from_dot_data
from sklearn.tree import export_graphviz
from sklearn ... | code |
17137542/cell_18 | [
"image_output_1.png"
] | from sklearn import manifold
from sklearn.cluster import KMeans
import matplotlib.pyplot as plt
import pandas as pd
df = pd.read_csv('../input/whisky.csv')
dist = []
for i in range(2, 20):
km = KMeans(n_clusters=i, n_init=10, max_iter=500, random_state=0)
km.fit(df.iloc[:, 2:-3])
dist.append(km.inertia... | code |
17137542/cell_8 | [
"text_html_output_1.png"
] | import pandas as pd
df = pd.read_csv('../input/whisky.csv')
df.describe() | code |
17137542/cell_16 | [
"text_html_output_1.png"
] | from sklearn import manifold
from sklearn.cluster import KMeans
import matplotlib.pyplot as plt
import pandas as pd
df = pd.read_csv('../input/whisky.csv')
dist = []
for i in range(2, 20):
km = KMeans(n_clusters=i, n_init=10, max_iter=500, random_state=0)
km.fit(df.iloc[:, 2:-3])
dist.append(km.inertia... | code |
17137542/cell_14 | [
"text_plain_output_1.png"
] | from sklearn import manifold
from sklearn.cluster import KMeans
import matplotlib.pyplot as plt
import pandas as pd
df = pd.read_csv('../input/whisky.csv')
dist = []
for i in range(2, 20):
km = KMeans(n_clusters=i, n_init=10, max_iter=500, random_state=0)
km.fit(df.iloc[:, 2:-3])
dist.append(km.inertia... | code |
17137542/cell_10 | [
"text_plain_output_1.png"
] | from sklearn.cluster import KMeans
import matplotlib.pyplot as plt
import pandas as pd
df = pd.read_csv('../input/whisky.csv')
dist = []
for i in range(2, 20):
km = KMeans(n_clusters=i, n_init=10, max_iter=500, random_state=0)
km.fit(df.iloc[:, 2:-3])
dist.append(km.inertia_)
plt.plot(range(2, 20), dist... | code |
17137542/cell_12 | [
"text_html_output_1.png"
] | from sklearn.cluster import KMeans
import matplotlib.pyplot as plt
import pandas as pd
df = pd.read_csv('../input/whisky.csv')
dist = []
for i in range(2, 20):
km = KMeans(n_clusters=i, n_init=10, max_iter=500, random_state=0)
km.fit(df.iloc[:, 2:-3])
dist.append(km.inertia_)
km = KMeans(n_clusters=5, ... | code |
17144010/cell_13 | [
"text_plain_output_1.png"
] | import keras as K
import numpy as np
import tensorflow as tf
np.random.seed(1)
tf.set_random_seed(1)
tf.logging.set_verbosity(tf.logging.ERROR)
discriminator = K.Sequential()
depth = 64
dropout = 0.4
input_shape = (28, 28, 1)
discriminator.add(K.layers.Conv2D(depth * 1, 5, strides=2, input_shape=input_shape, paddin... | code |
17144010/cell_11 | [
"application_vnd.jupyter.stderr_output_1.png"
] | import keras as K
import numpy as np
import tensorflow as tf
np.random.seed(1)
tf.set_random_seed(1)
tf.logging.set_verbosity(tf.logging.ERROR)
discriminator = K.Sequential()
depth = 64
dropout = 0.4
input_shape = (28, 28, 1)
discriminator.add(K.layers.Conv2D(depth * 1, 5, strides=2, input_shape=input_shape, paddin... | code |
17144010/cell_8 | [
"text_plain_output_1.png"
] | from tensorflow.examples.tutorials.mnist import input_data
import numpy as np
import tensorflow as tf
np.random.seed(1)
tf.set_random_seed(1)
from tensorflow.examples.tutorials.mnist import input_data
x_train = input_data.read_data_sets('mnist', one_hot=True).train.images
x_train = x_train.reshape(-1, 28, 28, 1).as... | code |
17144010/cell_3 | [
"text_plain_output_1.png"
] | import numpy as np
import keras as K
import tensorflow as tf
import pandas as pd
import os
from matplotlib import pyplot as plt
import seaborn as sns
os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2' | code |
73067082/cell_1 | [
"text_plain_output_1.png"
] | import os
import numpy as np
import pandas as pd
from sklearn.ensemble import RandomForestRegressor
from sklearn.metrics import mean_absolute_error
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import LabelEncoder
from xgboost import XGBRegressor
import os
for dirname, _, filenames in... | code |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.