0% found this document useful (0 votes)
12 views

Da Program

Uploaded by

alishacalista238
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
12 views

Da Program

Uploaded by

alishacalista238
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 18

PROGRAM - 1

To get the input from user and perform numerical operations (MAX, MIN, AVG, SUM, SQRT,
ROUND)using in Python.

def get_user_input():
nums = []
while True:
num = input("Enter a number (or 'done' to finish): ")
if num.lower() == 'done':
break
else:
try:
nums.append(float(num))
except ValueError:
print("Invalid input. Please enter a valid number.")
return nums

def numerical_operations(nums):
if not nums:
print("No numbers entered.")
return

maximum = max(nums)
minimum = min(nums)
average = sum(nums) / len(nums)
total_sum = sum(nums)
sqrt_values = [round(num ** 0.5, 2) for num in nums]
rounded_values = [round(num) for num in nums]

print("Maximum:", maximum)
print("Minimum:", minimum)
print("Average:", average)
print("Sum:", total_sum)
print("Square roots:", sqrt_values)
print("Rounded values:", rounded_values)

def main():
print("Enter numbers to perform numerical operations.")
numbers = get_user_input()
numerical_operations(numbers)

if __name__ == "__main__":
main()
OUTPUT

Enter numbers to perform numerical operations.


Enter a number (or 'done' to finish): 10
Enter a number (or 'done' to finish): 20
Enter a number (or 'done' to finish): 30
Enter a number (or 'done' to finish): done
Maximum: 30.0
Minimum: 10.0
Average: 20.0
Sum: 60.0
Square roots: [3.16, 4.47, 5.48]
Rounded values: [10, 20, 30]
PROGRAM - 5

To perform dimensionality reduction operation using PCA for Houses Data Set in Python.

import pandas as pd
from sklearn.decomposition import PCA
from sklearn.preprocessing import StandardScaler

def apply_pca(data, num_components=2):


numeric_data = data.select_dtypes(include=['int64', 'float64'])
scaler = StandardScaler()
scaled_data = scaler.fit_transform(numeric_data)
pca = PCA(n_components=num_components)
pca_data = pca.fit_transform(scaled_data)
pca_columns = [f'PC{i+1}' for i in range(num_components)]
pca_df = pd.DataFrame(data=pca_data, columns=pca_columns)
final_data = pd.concat([pca_df, data], axis=1)
return final_data

url =
"https://raw.githubusercontent.com/dipanjanS/practical-machine-learning-with-python/master
/notebooks/Ch05_Machine_Learning_Pipeline/AB_NYC_2019.csv"
data = pd.read_csv(url)

result = apply_pca(data)
print(result.head())
OUTPUT
PROGRAM - 6

To Perform Simple Logistic Regression with Python.

import numpy as np
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.linear_model import LogisticRegression
from sklearn import datasets
from sklearn.metrics import accuracy_score, confusion_matrix

iris = datasets.load_iris()
X = iris.data
y = iris.target

X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)

model = LogisticRegression()

model.fit(X_train, y_train)

y_pred = model.predict(X_test)

accuracy = accuracy_score(y_test, y_pred)


print("Accuracy:", accuracy)

conf_matrix = confusion_matrix(y_test, y_pred)


print("Confusion Matrix:")
print(conf_matrix)

OUTPUT

Accuracy: 1.0
Confusion Matrix:
[[10 0 0]
[ 0 9 0]
[ 0 0 11]]
PROGRAM - 7

To perform data import/export (.CSV, .XLS, .TXT) operations using data frames in Python.
import pandas as pd .

import pandas as pd

data = {
'Name': ['John', 'Anna', 'Peter', 'Linda'],
'Age': [28, 35, 42, 32],
'City': ['New York', 'Paris', 'Berlin', 'London']
}

df = pd.DataFrame(data)

df.to_csv('sample_data.csv', index=False)
df.to_excel('sample_data.xlsx', index=False)
df.to_csv('sample_data.txt', sep='\t', index=False)

df_csv = pd.read_csv('sample_data.csv')
df_excel = pd.read_excel('sample_data.xlsx')
df_txt = pd.read_csv('sample_data.txt', sep='\t')

print("Data imported from CSV file:")


print(df_csv)
print("\nData imported from Excel file:")
print(df_excel)
print("\nData imported from text file:")
print(df_txt)
OUTPUT
PROGRAM - 8

To perform K-Means clustering operation and visualize for iris data set in Python.

import pandas as pd
import numpy as np
import seaborn as sns
import matplotlib.pyplot as plt
from sklearn import tree

from sklearn.metrics import accuracy_score, classification_report


from sklearn.datasets import load_iris
from sklearn.tree import DecisionTreeClassifier
from sklearn.model_selection import train_test_split

import warnings
warnings.filterwarnings('ignore')

iris = load_iris()
iris = sns.load_dataset('iris')
iris.head()

iris_setosa = iris.loc[iris["species"] == "Iris-setosa"]


iris_virginica = iris.loc[iris["species"] == "Iris-virginica"]
iris_versicolor = iris.loc[iris["species"] == "Iris-versicolor"]

sns.FacetGrid(iris, hue="species", size=3).map(sns.distplot, "petal_length").add_legend()


sns.FacetGrid(iris, hue="species", size=3).map(sns.distplot, "petal_width").add_legend()
sns.FacetGrid(iris, hue="species", size=3).map(sns.distplot, "sepal_length").add_legend()
plt.show()
OUTPUT
PROGRAM - 9

Write R/Python script to diagnose any disease using KNN classification and plot the resultsin
Python.
PROGRAM - 10

To perform market basket analysis using Association Rules (Apriori) in Python.

from mlxtend.frequent_patterns import apriori


from mlxtend.frequent_patterns import association_rules
import pandas as pd

data = {'TID': [1, 2, 3, 4, 5],


'Items': [['Milk', 'Bread', 'Butter'],
['Milk', 'Bread', 'Eggs'],
['Milk', 'Bread', 'Butter', 'Eggs'],
['Milk', 'Bread'],
['Milk', 'Eggs']]}

df = pd.DataFrame(data)
df['Items'] = df['Items'].apply(lambda x: ','.join(x))

one_hot_encoded = df['Items'].str.get_dummies(sep=',')

frequent_itemsets = apriori(one_hot_encoded, min_support=0.5, use_colnames=True)

rules = association_rules(frequent_itemsets, metric="confidence", min_threshold=0.5)

print("Frequent Itemsets:")
print(frequent_itemsets)

print("\nAssociation Rules:")
print(rules)
OUTPUT
PROGRAM - 11

To perform data pre-processing operations:


i) Handling Missing data
ii) Min-Max normalization

import pandas as pd
from sklearn.preprocessing import MinMaxScaler

data = {
'A': [1, 2, None, 4, 5],
'B': [10, 20, 30, None, 50],
'C': [100, 200, 300, 400, None]
}

df = pd.DataFrame(data)

df.fillna(df.mean(), inplace=True)

scaler = MinMaxScaler()
df_normalized = pd.DataFrame(scaler.fit_transform(df), columns=df.columns)

print("Original DataFrame with missing values:")


print(df)

print("\nDataFrame after handling missing values (filled with mean):")


print(df)

print("\nDataFrame after Min-Max normalization:")


print(df_normalized)

You might also like