Program
Program
same.
Program Code :
importnumpy as np
import pandas as pd
data=pd.read_csv(r"C:\Users\jackw\Downloads\iris.csv")
iris=datasets.load_iris()
data=pd.DataFrame(data=np.c_[iris['data'],iris['target']],
columns=iris['feature_names']+['target'])
data['target_names']=data['target'].replace(dict(enumerate(iris.target_names)))
print(data.head())
data.info()
print(data.describe())
Label_encoder=LabelEncoder()
data['target_names']=Label_encoder.fit_transform(data['target_names'])
print(data['target_names'])
print(Label_encoder.classes_)
print(data['target_names'].value_counts())
One_hot=OneHotEncoder()
transformed_data=One_hot.fit_transform(data['target_names'].values.reshape(-1,1)
).toarray()
print(One_hot.categories_)
transformed_data=pd.DataFrame(transformed_data,
columns=['setosa','versicolor','virginica'])
print(transformed_data.head())
2. Write a program to visualize the iris dataset using sepal and petal values.
Program Code:
importnumpy as np
import pandas as pd
importmatplotlib.pyplot as plt
iris = datasets.load_iris()
print(data)
ax.set(xlabel=iris.feature_names[0], ylabel=iris.feature_names[1])
plt.show()
fig, ax = plt.subplots()
ax.set(xlabel=iris.feature_names[2], ylabel=iris.feature_names[3])
plt.show()
3. Write a program to apply linear regression on the iris dataset and find the
error of prediction.
import pandas as pd
importmatplotlib.pyplot as plt
iris = datasets.load_iris()
X = iris.data
y = iris.target
model = LinearRegression()
model.fit(X_train, y_train)
y_pred = model.predict(X_test)
import pandas as pd
iris = datasets.load_iris()
X = data
y = iris.target
logreg.fit(X_train, y_train)
y_pred = logreg.predict(X_test)
import pandas as pd
iris = datasets.load_iris()
X = iris.data
y = iris.target
model = GaussianNB()
model.fit(X_train, y_train)
y_pred = model.predict(X_test)
accuray = accuracy_score(y_pred, y_test)
import pandas as pd
importmatplotlib.pyplot as plt
iris = datasets.load_iris()
X = iris.data
y = iris.target
y_kmeans = kmeans.fit_predict(X)
'Centroids')
plt.legend()
plt.show()
Part B
import pandas as pd
import warnings
iris = datasets.load_iris()
print('Average of Feature')
print(data.mean())
print('\nFeature Variance')
print(data.var())
scaler = StandardScaler()
scaler.fit(data)
iris_scaled_std = scaler.transform(data)
iris_df_std_scaled = pd.DataFrame(data=iris_scaled_std,
columns=iris.feature_names)
print(iris_df_std_scaled.mean())
print(iris_df_std_scaled.var())
scaler = MinMaxScaler()
scaler.fit(data)
iris_scaled_minmax = scaler.transform(data)
iris_df_minmax_scaled = pd.DataFrame(data=iris_scaled_minmax,
columns=iris.feature_names)
print(iris_df_minmax_scaled.mean())
print(iris_df_minmax_scaled.var())
print(iris_df_minmax_scaled.min())
dataset.
Program Code:
importnumpy as np
import pandas as pd
fromsklearn.preprocessing import KBinsDiscretizer
import warnings
iris = datasets.load_iris()
print(data)
print(imputer)
imputer = imputer.fit(data)
imputed_dataset = imputer.transform(data)
print(imputed_dataset)
quantile_transformed_data = iris_quantile_transform.fit_transform(data)
print(pd.DataFrame(quantile_transformed_data))
uniform_transformed_data = iris_uniform_transform.fit_transform(data)
print(pd.DataFrame(uniform_transformed_data))
Program Code:
importnumpy as np
import pandas as pd
importmatplotlib.pyplot as plt
import mpl_toolkits.mplot3d
iris = datasets.load_iris()
X = data
y = iris.target
print(data)
plt.clf()
ax = fig.add_subplot(111, projection="3d", elev=48, azim=134)
plt.cla()
pca = decomposition.PCA(n_components=3)
pca.fit(X)
X = pca.transform(X)
print(X)
ax.text3D(
name,
horizontalalignment="center",
ax.yaxis.set_ticklabels([])
ax.zaxis.set_ticklabels([])
plt.show()
import pandas as pd
iris = datasets.load_iris()
X = iris.data
y = iris.target
SVM_classifier.fit(X_train, y_train)
y_pred = SVM_classifier.predict(X_test)
print(SVM_cm)
5.Develop Decision Tree Classification model for a given dataset and use it to
classify a
new sample.
iris = datasets.load_iris()
X = iris.data
y = iris.target
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3,
random_state=42)
dt_classifier = DecisionTreeClassifier(random_state=42)
dt_classifier.fit(X_train, y_train)
y_pred = dt_classifier.predict(X_test)
print("\nConfusion Matrix:")
print(conf_matrix)
print("\nClassification Report:")
print(class_report)
new_sample = [[5.1, 3.5, 1.4, 0.2]] # Example: a new Iris sample (sepal_length,
sepal_width,
petal_length, petal_width)
new_prediction = dt_classifier.predict(new_sample)
import numpy as np
from sklearn import datasets
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import StandardScaler
from sklearn.neighbors import KNeighborsClassifier
from sklearn.metrics import accuracy_score, confusion_matrix,
classification_report
iris = datasets.load_iris()
X = iris.data
y = iris.target
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3,
random_state=42)
scaler = StandardScaler()
X_train = scaler.fit_transform(X_train)
X_test = scaler.transform(X_test)
knn = KNeighborsClassifier(n_neighbors=3)
knn.fit(X_train, y_train)
y_pred = knn.predict(X_test)
accuracy = accuracy_score(y_test, y_pred)
print("\nConfusion Matrix:")
print(conf_matrix)
print("\nClassification Report:")
print(class_report)