0% found this document useful (0 votes)
102 views23 pages

Minor Project Report

This project aims to develop a machine learning model to predict house prices in Bengaluru using a dataset of property details. Linear regression is used as the primary algorithm for price prediction, validated with cross-validation and optimized with GridSearchCV. The goal is to provide accurate predictions to help stakeholders make informed decisions in the real estate market.

Uploaded by

jaiswalsamrat12
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
102 views23 pages

Minor Project Report

This project aims to develop a machine learning model to predict house prices in Bengaluru using a dataset of property details. Linear regression is used as the primary algorithm for price prediction, validated with cross-validation and optimized with GridSearchCV. The goal is to provide accurate predictions to help stakeholders make informed decisions in the real estate market.

Uploaded by

jaiswalsamrat12
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
You are on page 1/ 23

A PROJECT REPORT

on

“HOUSE PRICE PREDICTION”

Submitted to
KIIT Deemed to be University

In Partial Fulfilment of the Requirement for the Award of

BACHELOR’S DEGREE IN
INFORMATION TECHNOLOGY

BY

OM PATEL 21052775
AMULYA JAISWAL 21052732
DEEPAK SINGH 21052752
DHANJEE TIWARI 21052755
NITIN RAJ 21052773

UNDER THE GUIDANCE OF


MR. ABHISHEK RAJ

SCHOOL OF COMPUTER ENGINEERING


KALINGA INSTITUTE OF INDUSTRIAL TECHNOLOGY
BHUBANESWAR, ODISHA - 751024
KIIT Deemed to be University
School of Computer Engineering
Bhubaneswar, ODISHA 751024

CERTIFICATE
This is to certify that the project entitled
“HOUSE PRICE PREDICTION “
submitted by

OM PATEL 21052775
AMULYA JAISWAL 21052732
DEEPAK SINGH 21052752
DHANJEE TIWARI 21052755
NITIN RAJ 21052773

It is a record of bonafide work carried out by them, in the partial fulfilment of the
requirement for the award of Degree of Bachelor of Engineering (Computer Sci-
ence & Engineering OR Information Technology) at KIIT Deemed to be university,
Bhubaneswar. This work is done during year 2022-2023, under our guidance.

Date: / /

(MR. ABHISHEK RAJ)


Project Guide
Acknowledgements

We are profoundly grateful to ABHISHEK RAJ of KIIT for his expert guidance
and continuous encouragement throughout to see that this project rights its target
since its commencement to its completion.
ABSTRACT

This project aims to utilize data analysis and machine learning techniques to
develop a predictive model for real estate prices in Bengaluru, India. The
primary objective is to facilitate informed decision-making in the housing market
by accurately forecasting property prices. The dataset used in this project
comprises comprehensive details of residential properties in Bengaluru,
including key features such as location, size, total square feet area, number of
bathrooms, and price. Through meticulous data preprocessing, outlier removal,
and feature engineering, the dataset is prepared for modeling.

Linear Regression is employed as the primary algorithm for price prediction,


validated using cross-validation techniques, and further optimized using
GridSearchCV for enhanced model performance. This methodology ensures the
development of a robust predictive model capable of providing valuable insights
into the Bengaluru real estate market. The project's contribution lies in its ability
to offer accurate price predictions, thereby empowering stakeholders, including
buyers, sellers, and investors, to make well-informed decisions. Additionally, it
demonstrates proficiency in data analysis and machine learning techniques,
underscoring the potential for leveraging advanced analytical tools to address
complex challenges in the housing sector.

Overall, this project serves as a testament to the efficacy of data-driven


approaches in transforming traditional industries like real estate, enabling
stakeholders to navigate the market landscape with greater confidence and
precision.

Keywords: Real estate, Bengaluru, Housing, Dataset, Preprocessing, Linear


Regression, Cross-validation, Outlier removal, Feature engineering, GridSearchCV.
Contents
1 Introduction 1

2 Basic Concepts/ Literature Review 2


2.1 Sub Section Name........................... 2

3 Problem Statement / Requirement Specifications 3


3.1 Project Planning........................... 3
3.2 Project Analysis (SRS)................. 3
3.3 System Design ………………….. 3
3.3.1 Design Constraints …… 3
3.3.2 System Architecture (UML) / Block Diagram … 3

4 Implementation 4
4.1 Methodology / Proposal ........................... 4
4.2 Testing / Verification Plan ……………. 4
4.3 Result Analysis / Screenshots …………. 4
4.4 Quality Assurance …………………….. 4

5 Standard Adopted 5
5.1 Design Standards . . . . . . . . . . . . . . . 5
5.2 Coding Standards . . . . . . . . . . . . . . 5
5.3 Testing Standards . . . . . . . . . . . . . . . 5

6 Conclusion and Future Scope 6


6.1 Conclusion ……………………….. 6
6.2 Future Scope ………………………. 6

References 7

Individual Contribution 8

Plagiarism Report 9
List of Figures

1.1 IMAGE CAPTION ......................... 2

4.1 IMAGE CAPTION ......................... 9


NAME OF PROJECT

Chapter 1

Introduction
The Bengaluru real estate market stands as a dynamic and ever-evolving
landscape, characterized by fluctuating property prices and diverse consumer
preferences. In this context, the ability to accurately predict real estate prices
becomes paramount for stakeholders, including buyers, sellers, and investors, to
make informed decisions and navigate the market effectively. Leveraging the
power of data analysis and machine learning techniques, this project endeavors
to address this critical need by developing a predictive model for real estate
prices in Bengaluru.

At the core of this project lies a comprehensive dataset comprising detailed


information about residential properties in Bengaluru. This dataset encompasses
essential features such as location, size, total square feet area, number of
bathrooms, and price, providing a rich source of information for analysis and
modeling. Through meticulous data preprocessing, outlier removal, and feature
engineering, the dataset is refined and prepared for modeling, ensuring the
quality and integrity of the data.

The methodology employed in this project revolves around the utilization of


Linear Regression as the primary algorithm for price prediction. This modeling
approach is further augmented by the use of cross-validation techniques to
validate the model's performance and GridSearchCV for hyperparameter tuning,
thereby optimizing model accuracy and robustness. By adhering to best practices
in data science and machine learning, this project aims to develop a reliable and
accurate predictive model capable of providing valuable insights into the
Bengaluru real estate market.

Through the culmination of these efforts, this project seeks to contribute to the
advancement of decision-making processes in the Bengaluru real estate sector.
By harnessing the potential of data-driven insights, stakeholders can gain a
deeper understanding of market dynamics, mitigate risks, and capitalize on
opportunities, ultimately fostering a more transparent and efficient real estate
ecosystem.

School of Computer Engineering, KIIT, BBSR 1


NAME OF PROJECT
Chapter 2

Basic Concepts/ Literature Review


2.1 Machine Learning

Machine learning is a branch of artificial intelligence (AI) that enables


computers to learn from data and improve their performance over time without
being explicitly programmed. There are several types of machine learning
algorithms, including supervised learning, unsupervised learning, and
reinforcement learning.

2.1.1 Supervised Learning

Supervised learning is a type of machine learning where the algorithm learns


from labeled data, which consists of input-output pairs. The goal is to learn a
mapping function from the input variables to the output variable, allowing the
algorithm to make predictions on unseen data. In supervised learning, the
algorithm is trained on a dataset that includes both input features and
corresponding output labels. Common supervised learning algorithms include:

 Linear Regression: A regression algorithm used for predicting a


continuous target variable based on one or more input features. It models
the relationship between the independent variables (features) and the
dependent variable (target) using a linear equation.

 Logistic Regression: A classification algorithm used for predicting the


probability of occurrence of a categorical target variable. It models the
relationship between the independent variables and the probability of
belonging to a particular class using the logistic function.

 Decision Trees: A versatile algorithm that can be used for both


classification and regression tasks. It partitions the feature space into
regions and makes predictions based on the majority class or average
value of the target variable within each region.

 Support Vector Machines (SVM): A powerful algorithm for classification


and regression tasks that finds the optimal hyperplane to separate data
points into different classes or predict continuous values.

2.1.2 Unsupervised Learning


Unsupervised learning involves training algorithms on unlabeled data to uncover
hidden patterns or structures within the data. Unlike supervised learning, there
are no predefined output labels, and the algorithm must find meaningful
representations or groupings in the data. Common unsupervised learning
algorithms include:

K-means Clustering: A clustering algorithm that partitions data points into k


clusters based on similarity, with each cluster represented by the mean of the
data points assigned to it.

Hierarchical Clustering: A clustering algorithm that creates a tree-like hierarchy


of clusters by recursively merging or splitting clusters based on their similarity.

2.1.3 Cross-Validation

Cross-validation is a technique used to evaluate the performance of machine


learning models by partitioning the dataset into multiple subsets, or folds. The
model is trained on a subset of the data and tested on the remaining fold, and this
process is repeated multiple times to obtain an unbiased estimate of the model's
performance. Common cross-validation methods include k-fold cross-validation,
leave-one-out cross-validation, and stratified cross-validation.

2.1.4 Hyperparameter Tuning

Hyperparameter tuning involves selecting the optimal values for the


hyperparameters of a machine learning algorithm to maximize performance.
Hyperparameters are parameters that are set prior to training and cannot be
learned from the data. Techniques such as grid search and random search are
used to systematically explore the hyperparameter space and identify the best
combination of values for optimal model performance.

2.1.5 Model Evaluation Metrics

Model evaluation metrics are used to assess the performance of machine learning
models and compare different algorithms. Common evaluation metrics for
regression tasks include mean squared error (MSE), root mean squared error
(RMSE), mean absolute error (MAE), and R-squared (coefficient of
determination). For classification tasks, metrics such as accuracy, precision,
recall, F1-score, and area under the ROC curve (AUC-ROC) are commonly
used. These metrics provide valuable insights into the model's predictive
accuracy, robustness, and generalization capabilities.

School of Computer Engineering, KIIT, BBSR 2


NAME OF PROJECT

Chapter 3

Problem Statement / Requirement


Specifications

The Bengaluru real estate market presents challenges for stakeholders due to
its dynamic nature and fluctuating property prices. In this context, the lack of
accurate predictive models hampers informed decision-making for buyers,
sellers, and investors. The problem statement for this project revolves around
the need to develop a reliable predictive model for real estate prices in
Bengaluru using data analysis and machine learning techniques.

The objective is to address these challenges by leveraging a comprehensive


dataset of residential properties in Bengaluru and applying advanced data
preprocessing, modeling, and validation techniques. The goal is to develop a
predictive model that provides actionable insights into the Bengaluru real
estate market, enabling stakeholders to make informed decisions and navigate
the market landscape with confidence.

3.1 Project Planning

 Requirement Gathering: Identify the key stakeholders and their


requirements for the predictive model. Gather input from real estate
professionals, buyers, sellers, and investors to understand their needs and
pain points. Define the scope of the project, including the target audience,
geographical focus (Bengaluru), and desired outcomes.

 Data Collection and Preparation: Collect a comprehensive dataset


containing details of residential properties in Bengaluru. Ensure the
dataset includes relevant features such as location, size, total square feet
area, number of bathrooms, and price. Cleanse the data by handling
missing values, removing duplicates, and addressing any inconsistencies.
Perform exploratory data analysis (EDA) to gain insights into the dataset's
structure and characteristics.

 Feature Engineering: Engineer new features or transform existing ones to


enhance the predictive power of the model. Explore techniques such as
one-hot encoding for categorical variables (e.g., location) and scaling for
numerical variables. Consider domain-specific knowledge to create
meaningful features that capture the nuances of the real estate market.

 Model Selection and Development: Choose appropriate machine learning


algorithms for price prediction, considering factors such as interpretability,
performance, and scalability. Develop and train the model using
techniques like Linear Regression, Decision Trees, or ensemble methods.
Implement cross-validation techniques to evaluate model performance and
ensure generalization to unseen data.

 Hyper parameter Tuning and Optimization: Fine-tune model hyper


parameters using techniques like GridSearchCV or
RandomizedSearchCV. Optimize model performance based on metrics
such as mean squared error (MSE) or R-squared. Iterate on the model
development process to refine and improve predictive accuracy.

 Model Evaluation and Validation: Evaluate the trained model using


appropriate metrics and validation techniques. Assess model performance
on test data to ensure its reliability and generalization. Validate the
model's predictions against real-world scenarios and compare them with
industry benchmarks.

 Deployment and Monitoring: Deploy the trained model in a production


environment for real-time predictions. Implement monitoring mechanisms
to track model performance and detect drift or degradation over time.
Continuously update and retrain the model using new data to maintain its
relevance and accuracy.

 Documentation and Reporting: Document the entire project lifecycle,


including data sources, methodology, and model implementation details.
Prepare comprehensive reports or presentations summarizing key findings,
insights, and recommendations for stakeholders. Ensure transparency and
reproducibility by documenting code, configurations, and assumptions
made during the project.

3.2 Project Analysis

 Requirement Analysis: Review the gathered requirements or problem


statement to ensure clarity and completeness. Identify any ambiguities,
inconsistencies, or gaps in the requirements that may impact the project's
success. Seek clarification from stakeholders or domain experts to resolve
any uncertainties or misunderstandings.

 Feasibility Assessment: Evaluate the feasibility of the project in terms of


technical, financial, and resource constraints. Assess the availability of
data sources, computing infrastructure, and expertise required for model
development. Consider potential risks and challenges that may arise
during the project lifecycle and develop mitigation strategies.
 Stakeholder Alignment: Engage with key stakeholders to align
expectations and ensure their buy-in for the project objectives and
approach. Communicate the project scope, timelines, and deliverables to
stakeholders to manage expectations effectively. Address any concerns or
objections raised by stakeholders and incorporate feedback into the project
plan.

 Technology and Tools Selection: Evaluate available technologies and


tools for data preprocessing, modeling, and deployment. Choose
appropriate programming languages, libraries, and frameworks based on
factors such as scalability, ease of use, and community support. Consider
compatibility with existing systems and infrastructure within the
organization.

 Risk Identification and Mitigation: Identify potential risks and


uncertainties that may impact project success, such as data quality issues,
algorithmic complexity, or regulatory compliance. Develop a risk
management plan outlining strategies to mitigate, transfer, or accept
identified risks. Establish contingency plans and alternative approaches to
address unforeseen challenges or obstacles encountered during project
execution.

 Validation and Verification: Validate the project requirements against


stakeholder expectations and business objectives. Verify that the proposed
solution aligns with the identified problem statement and addresses the
underlying business needs effectively. Conduct peer reviews or
walkthroughs to validate project documentation, design, and
implementation decisions.

 Continuous Improvement: Foster a culture of continuous improvement by


soliciting feedback from team members and stakeholders throughout the
project lifecycle. Identify lessons learned and best practices to inform
future projects and enhance organizational capabilities. Monitor project
progress and performance metrics to track success criteria and identify
areas for optimization or refinement.

3.3 System Design

3.3.1 Design Constraints

 Software Environment: Utilization of data analysis and machine learning


libraries such as pandas, NumPy, scikit-learn, and matplotlib for data
processing, modeling, and visualization. Implementation of programming
languages like Python for algorithm development and model deployment.
Usage of development environments such as Jupyter Notebook or Google
Colab for interactive development and experimentation.
 Hardware Environment: Requirement for computational resources capable
of handling large datasets and complex machine learning algorithms.
Availability of sufficient memory (RAM) and processing power
(CPU/GPU) to train and evaluate machine learning models efficiently.
Consideration of cloud-based solutions like AWS, Google Cloud Platform,
or Azure for scalable computing resources if local hardware constraints
exist.

 Experimental Setup: Access to a diverse and representative dataset of


residential properties in Bengaluru, including relevant features such as
location, size, total square feet area, number of bathrooms, and price.
Establishment of a data preprocessing pipeline to clean, transform, and
prepare the dataset for modeling, including handling missing values,
outlier detection, and feature engineering. Implementation of cross-
validation techniques to evaluate model performance and ensure
robustness to variations in the dataset. Utilization of GridSearchCV or
similar hyperparameter tuning methods to optimize model parameters and
improve predictive accuracy.

 Environmental Constraints: Consideration of regulatory and ethical


constraints related to data privacy and security when handling sensitive
information such as property details and transaction records. Adherence to
industry standards and best practices for data handling, modeling, and
deployment to ensure compliance with legal requirements and industry
norms. Awareness of environmental factors such as seasonal variations,
market trends, and economic indicators that may influence real estate
prices in Bengaluru.

3.3.2 System Architecture OR Block Diagram

The project's system architecture is designed to encompass the end-to-end


process of developing and deploying a predictive model for real estate price
prediction in Bengaluru. The architecture consists of several key components,
each serving a specific purpose in the overall workflow. Below is an overview of
the system architecture:

 Data Collection and Preprocessing: Data collection involves gathering a


comprehensive dataset containing details of residential properties in
Bengaluru, including features such as location, size, total square feet area,
number of bathrooms, and price. Data preprocessing encompasses
cleaning, transforming, and preparing the dataset for modeling. This
includes handling missing values, removing duplicates, and performing
feature engineering to extract meaningful insights from the data.

 Model Development: Model development involves selecting appropriate


machine learning algorithms, such as Linear Regression, Decision Trees,
or ensemble methods, for price prediction. The chosen algorithms are
trained on the preprocessed dataset using techniques like cross-validation
to evaluate model performance and ensure generalization to unseen data.
 Hyperparameter Tuning and Optimization: Hyperparameter tuning is
performed using techniques like GridSearchCV or RandomizedSearchCV
to fine-tune model parameters and optimize predictive accuracy. Model
optimization aims to improve performance metrics such as mean squared
error (MSE) or R-squared, enhancing the model's ability to make accurate
predictions.

 Model Deployment: The trained and optimized model is deployed in a


production environment for real-time predictions. Deployment may
involve integrating the model into existing systems or applications using
APIs or web services to enable seamless interaction with end-users.

 Monitoring and Maintenance: Continuous monitoring mechanisms are


implemented to track model performance and detect any drift or
degradation over time. Regular maintenance and updates are performed to
ensure the model remains relevant and effective in capturing changes in
the real estate market.

School of Computer Engineering, KIIT, BBSR 3

NAME OF PROJECT

Chapter 4

Implementation
In this section, present the implementation done by you during the project
development.

4.1 Methodology OR Proposal

 Data Collection: Gathered a comprehensive dataset containing details of


residential properties in Bengaluru from reliable sources such as real estate
websites, property listings, and government databases. The dataset
includes features such as location, size, total square feet area, number of
bathrooms, and price.

 Data Preprocessing: Conducted data preprocessing to clean, transform,


and prepare the dataset for modeling. Handled missing values, removed
duplicates, and addressed any inconsistencies in the data. Performed
feature engineering to extract relevant information and enhance the
predictive power of the model.

 Model Development: Selected Linear Regression as the primary algorithm


for price prediction due to its simplicity and interpretability. Split the
dataset into training and testing sets to evaluate model performance.
Trained the Linear Regression model on the training data using techniques
like cross-validation to assess its generalization to unseen data.

 Hyperparameter Tuning and Optimization: Fine-tuned model


hyperparameters using GridSearchCV to optimize predictive accuracy.
Explored different combinations of hyperparameters to identify the
optimal configuration for the Linear Regression model.

 Model Evaluation: Evaluated the trained model's performance using


metrics such as mean squared error (MSE), R-squared, and root mean
squared error (RMSE).Compared model predictions against actual prices
to assess accuracy and reliability.

 Model Deployment: Deployed the trained Linear Regression model in a


production environment for real-time predictions. Implemented a web
service or API to enable seamless integration with existing systems or
applications.

 Monitoring and Maintenance: Established monitoring mechanisms to track


model performance and detect any drift or degradation over time.
Scheduled regular maintenance and updates to ensure the model remains
relevant and effective in capturing changes in the real estate market.

4.2 Verification Plan

Test Test Case Title Test Condition System Behavior Expected Result

ID
Preprocessed data is
Data System clean, transformed,
Preprocessing Preprocessed preprocesses data and ready for
T01 Test data is available successfully modeling.
Model is trained System trains the Trained model is
Model Training on the training model without ready for evaluation
T02 Test dataset errors and deployment.
System evaluates Model performance
Model model performance metrics meet
performance using metrics such predefined criteria for
Model Evaluation metrics are as MSE, R-squared, accuracy and
T03 Test calculated and RMSE reliability.
Model is
deployed in a System deploys the Deployed model is
Model production model without accessible for real-
T04 Deployment Test environment errors time predictions.
Monitoring alerts are
System monitors generated when
Monitoring model performance model performance
Monitoring and mechanisms are and detects drift or deviates from
T05 Maintenance Test implemented degradation expected behavior.

4.3 Result Analysis OR Screenshots


In this subsection, the output of the experiment or study in terms of some graphs,
plots must be presented. Also, if some implementation is done then it’s
screenshots can be presented here, so as to showcase the proof of the output.

4.4 Quality Assurance


In the working organization, if some department is there to verify the quality of
your work, they can produce a certificate or guidelines followed.

School of Computer Engineering, KIIT, BBSR 4


NAME OF PROJECT

Chapter 5

Standards Adopted

5.1 Design Standards


In all the engineering streams, there are predefined design standards are present
such as IEEE, ISO etc. List all the recommended practices for project design. In
software the UML diagrams or database design standards also can be followed.

5.2 Coding Standards


Coding standards are collections of coding rules, guidelines, and best practices.
Few of the coding standards are:
1. Write as few lines as possible.
2. Use appropriate naming conventions.
3. Segment blocks of code in the same section into paragraphs.
4. Use indentation to marks the beginning and end of control structures. Clearly
specify the code between them.
5. Don’t use lengthy functions. Ideally, a single function should carry out a
single task.
…...

5.3 Testing Standards


There are some ISO and IEEE standards for quality assurance and testing of the
product. Mention the standards followed for testing and verification of your
project work.

School of Computer Engineering, KIIT, BBSR 5


NAME OF PROJECT

Chapter 6

Conclusion and Future Scope

6.1 Conclusion

In conclusion, the project aimed to develop a predictive model for


real estate price prediction in Bengaluru using data analysis and
machine learning techniques. Through the systematic methodology
outlined, including data collection, preprocessing, model
development, and deployment, significant progress was made
towards achieving this goal.

Key findings from the project include the successful preprocessing


of the dataset to clean and transform the data, the development of a
robust predictive model using Linear Regression, and the
deployment of the trained model for real-time predictions. The
project also demonstrated the importance of hyperparameter tuning
and optimization in improving model performance, as well as the
necessity of monitoring and maintenance to ensure the model's
continued relevance and effectiveness.

While the project has made significant strides in addressing the


problem statement and meeting the project objectives, there are
opportunities for further enhancements and refinements. Future
work could focus on incorporating additional features, exploring
alternative algorithms, or expanding the geographical scope beyond
Bengaluru to enhance the model's predictive accuracy and
generalization.

Overall, the project has contributed valuable insights and knowledge


to the field of real estate analytics and machine learning, providing
stakeholders with a powerful tool for informed decision-making in
the Bengaluru real estate market. Through collaboration, innovation,
and continuous improvement, the project has laid the foundation for
future advancements and applications in this domain.

6.2 Future Scope

1. Geographic Expansion: Extend the predictive model to cover


other cities or regions beyond Bengaluru, allowing for broader
market insights and predictions.

2. Feature Enhancement: Incorporate additional features such as


property amenities, neighborhood characteristics,
transportation accessibility, and economic indicators to
improve model accuracy and relevance.

3. Advanced Algorithms: Explore advanced machine learning


algorithms and techniques such as ensemble methods, neural
networks, and deep learning architectures to enhance predictive
performance and capture complex patterns in real estate data.

4. Dynamic Pricing Models: Develop dynamic pricing models


that adapt to changing market conditions, seasonal trends, and
fluctuations in demand and supply, providing more accurate
and timely predictions for property prices.

5. User Interface and Visualization: Build intuitive user interfaces


and visualization tools to enable stakeholders to interact with
the predictive model, explore data trends, and gain actionable
insights for decision-making.

6. Integration with Real Estate Platforms: Integrate the predictive


model with real estate platforms, property listing websites, and
mobile applications to offer value-added services such as price
recommendations, property valuation, and investment analysis
to users.

7. Automated Valuation Models (AVMs): Develop automated


valuation models (AVMs) for real-time property valuation,
leveraging machine learning algorithms and big data analytics
to streamline the appraisal process and improve accuracy.
School of Computer Engineering, KIIT, BBSR 6
NAME OF PROJECT

References
[1 ]S. M. Metev and V. P. Veiko, Laser Assisted Microtechnology, 2nd ed., R. M. Osgood, Jr., Ed. Berlin,
Germany: Springer-Verlag, 1998.

[2] Breckling, Ed., The Analysis of Directional Time Series: Applications to Wind Speed and Direction, ser.
Lecture Notes in Statistics. Berlin, Germany: Springer, 1989, vol. 61.

[3] S. Zhang, C. Zhu, J. K. O. Sin, and P. K. T. Mok, “A novel ultrathin elevated channel low-temperature poly-
Si TFT,” IEEE Electron Device Lett., vol. 20, pp. 569–571, Nov. 1999.

[4] M. Wegmuller, J. P. von der Weid, P. Oberson, and N. Gisin, “High resolution fiber distributed measurements
with coherent OFDR,” in Proc. ECOC’00, 2000, paper 11.3.4, p. 109.

[5] R. E. Sorace, V. S. Reinhardt, and S. A. Vaughn, “High-speed digital-to-RF converter,” U.S. Patent 5 668
842, Sept. 16, 1997.

[6] (2002) The IEEE website. [Online]. Available: http://www.ieee.org/

[7] M. Shell. (2002) IEEEtran homepage on CTAN. [Online]. Available:


http://www.ctan.org/tex-archive/macros/latex/contrib/supported/IEEEtran/

School of Computer Engineering, KIIT, BBSR 7


NAME OF PROJECT

SAMPLE INDIVIDUAL CONTRIBUTION REPORT:

<TITLE OF THE PROJECT IN FONT SIZE 14, FONT STYLE TIMES


NEW ROMAN, BOLD AND CENTERED>

<Student Name (in capital letters in font size 12, Times New Roman and
centered>
<Student Roll number (font size 12, Times New Roman and centered>

Abstract: A short description of the aim and objective of the project work carried out in 3-4
lines. This part should be common to all students in the group. The font size and style will
remain same from this point onwards. The font size will be 12 and font style will be Times
New Roman. The line spacing will be 1.5.
This report should be prepared in A4 page format with ‘default’ option under ‘Margin’ of the
‘Page Layout’ tab in Microsoft Word. Word limit for this section is 80.

Individual contribution and findings: The student should clearly indicate his/her role
in the project group and the contribution in implementing the project work. The student should
also outline his /her planning involved in implementing his/her part in the work. This
contribution report should be different for every student in the group. The student would also
write his./her technical findings and experience while implementing the corresponding part of
the project. The overall contribution report should not be less than 1 page for each student.
The Student should provide both the soft copy and signed hard copy to the project supervisor.

Individual contribution to project report preparation: Student should mention


his/her role in preparing the group project report indicating which chapter and portions
contributed.

Individual contribution for project presentation and demonstration: Student


should mention his/her role in preparing presentations and part of the project demonstrated.

Full Signature of Supervisor: Full signature of the


student:
……………………………. ……………………………..

School of Computer Engineering, KIIT, BBSR 8


TURNITIN PLAGIARISM REPORT
(This report is mandatory for all the projects and plagiarism
must be below 25%)

School of Computer Engineering, KIIT, BBSR 9

You might also like