Vehicle Management System Project using Python and SQLite

 

The complete development of this Vehicle Management System project using Django as the backend and sqlite3 as the database. This Python project has main sections Login/Signup, Dashboard, Vehicle, Driver, Booking, Repair, and Report which are explained in the coming slides in detail.

Functions Below:

Login/Signup

Users can signup/log in to the portal with this page. It takes in the necessary fields required for the user details

Dashboard

This is the dashboard in which you can view your details is shown and the user can edit the details.

Driver

Driver Section has two pages, one is to add a driver and the other is to view the list of drivers available. This section is only visible to the users which have admin access. The Driver list has the features to search and sort lists according to the fields.

Vehicles

In this section there are two pages one is to add vehicles and the other is to list the vehicles owned by the user. On the vehicle list page, the user can view individual details, edit the details and delete the vehicle.

Booking

The booking section has four pages one is the form to book a trip the second one is the success page where the booking details including the distance, cost, and duration are displayed using google maps API. On the success page, you have the option to pay which will take you to the payment page.

The payment page lets you enter the card details to pay. There is a page to display all the bookings made by the user wherein the user can search for the bookings. The admin will have another option which is to confirm the booking and a driver will be allotted and a mail will be sent to the user saying the mail is confirmed. The map option in the booking list page displays the route using google maps API. When the admin confirms the booking a confirmation mail will be sent to the user.

Repair

The repair section has two pages, one for reporting the issue and another page to show the issues made by the user. On the issues page, the admin has an option to solve the issue made by all the users.

Report

Each user can have a report of the trips that he has made. There is an option to mail the user the report for further use. Demo Here is the live demo of our project. The quality of the gif is a bit low.

Database tables:

Vehicle Table

Owner
Cost per KM
Price
Registration Plate
Vehicle Status
Insurance Status
Total KM traveled
Fuel Type
Mileage
Vehicle Type
Image

Report Table

Registered Date
Registered User
Vehicle Mileage
Issue

Driver Table

First Name
Last Name
National ID
Address
Email
Phone Number
License Category

BookingTable

Source
Destination
Distance
Booking Date
Start Date
End Date
Security Deposit
Allotted User
Allotted Driver
Vehicle
Cost
Duration

Corona Virus Prediction and Analysis Machine Learning Project

1.  Introduction 

Background 

Currently, there are many people, who are being affected by CoronaVirus. It started in China and now it is spreading all over the world. Till now, there is no medicine for this virus, and it’s killing millions of millions of people. So, it is a big question among all of us of how many people are going to be affected.

Problem Statement 

Currently, there is no application that can predict the spread of CoronaVirus for the future 30 days. So, with this project, we would like to create awareness among the people, by showing them how the corona rises for the future 30 days so that they can take some preventive measures by staying indoors.

Project Goal 

The main objective of this Corona Virus Prediction project is :

  • Future prediction of the increase/decrease in the number of active Coronavirus Cases for the next 30 days – for the whole world as well as for the United States of America. We have chosen the USA among all the counties as it is the highly affected country due to corona.
  • Future prediction of the increase/decrease in the number of deaths due to Coronavirus for the next 30 days.
  • Future prediction of the increase/decrease in the number of recovered cases due to Coronavirus for the next 30 days.

2.  Literature Review

 There is an outbreak of Corona in early December. This is caused due to severe acute respiratory syndrome coronavirus 2, which is basically the family of SARS virus. Many governments all over the world are issuing their own preventive measures to control the spread of coronavirus. So, we have conducted a literature review regarding this virus, based on the information that is publicly available.

Background of Literature Review:

China alerted WHO on 31st December 2019 that many people are reported to be suffering from Pneumonia, in Wuhan City. They reported that it started on Dec 8th, 2019, and there were an increasing number of patients who are working or living around the Huanan Seafood Wholesale Market.

When we started working on this project at the start of February, the Coronavirus was majorly prevalent in China. Initially, at the time of our project proposal, the mortality rate in China among all the confirmed cases is around 1.2% as of February 2020. And the mortality rate in all other countries, other than china was around only 0.2%. Among all the patients, who were admitted to the hospitals, the mortality rate, was around 11%. COVID-19 is increasing with great speed, and now there is a relatively very high mortality rate

A Way to Further Research :

So, we have performed this literature review, to analyze the spread of coronavirus. After analyzing how increasingly it’s spreading all over the world, we thought of performing our own prediction regarding this virus, so as to make people aware of its spread, and with this, they can take their own preventive measures, so that they do not fall prey to this dangerous virus.

We had very little amount of data when we started this project. It is a very trending topic all over the world. And millions of millions of people are losing their lives due to this virus. So, we are very curious to analyze this pandemic and so we have taken up this project.

We have found many datasets to collect data regarding the corona cases. Some of them include Kaggle, John Hoppkins, etc. So, we thought of choosing the dataset from John Hoppkins, as it’s updating the dataset on a daily basis. So, we collected the data and performed our own future predictions.

3. Methodology 

Approach

 So, basically, we have followed the below approach to kick-start our Corona Virus Prediction project:

  1. Firstly, we have started with research on choosing the datasets. On performing research on various datasets, we have finalized with John Hoppkins data set, as it gives us the live data on coronavirus.
  2. Secondly, we have collected the data and performed our preprocessing operation, so as to make our data ready for future predictions.
  3. Next, coming to choosing the machine learning algorithm. We have chosen appropriate machine learning(we will discuss below regarding this).
  4. Finally, we have performed our predictions to analyze the active cases, deaths, and recoveries for the next 30 days, based on the data available from the datasets and the chosen machine learning algorithm.

Figure: Approach

4.  Implications 

Benefits of the Project: 

  • This Corona Virus Prediction project helps in the prediction of coronavirus cases for the next 30 days, all over the world.
  • With this, we can also predict the increase in corona cases in the world.
  • By this, we can know how fast the coronavirus is spreading all over the world.
  • We can create awareness among people.
  • We can also create awareness in government so that they can take preventive measures to stop the spread of corona.

Lessons Learned:

Initially, I had no idea of a Machine learning algorithm. I started learning about machines from scratch. I bought some Udemy tutorials and through that, I learned everything step by step. At the start of the project, I am not even aware of what machine learning algorithm to use.

It was really an exciting experience doing this project. I am inspired to take up a Machine Learning Course for my next semester to learn deeply about Machine Learning Algorithms.

I tried my level best and contributed my 100% to this project.

Now, I came to know about machine learning, different types of machine learning Algorithms, and the differences between classification and regression algorithms -when to use what, creating test and train sets, building up the model, choosing the appropriate parameters, and performing future predictions. In the future, I would also love to take up a project related to Classification Algorithms.

5.  Conclusion 

  • Finally, to conclude, we have performed prediction using SVR and Polynomial Regression Algorithm.
  • SVR predictions are mainly for predicting the world case scenario, which includes confirmed, death, and recovered cases.
  • Polynomial Regression is used for the prediction of US Cases.
  • Based on the results, we believe that our predictions were almost accurate, with some little differences from the actual values.
  • This project can be further scalable, to include the predictions for various individual

6.  Appendix 

  • We have used Google Collab for our project. As we are two members of the team, we have chosen this, because it enables us to simultaneously work on the project from different
  • No Installation is Required.
  • We just need to have a google account. And we can easily create a Google Collaboratory file in our google drive, just like Google docs.
  • We will provide both .py files as well as .ipynb files along with this report, so as to run on google collab.
  • .ipynb can be uploaded to google collab directly and the results of the projects can be easily checked.

Smart Agriculture System Project on IBM IOT platform using NodeRed framework

Project Scope

We need to follow these steps to complete our Smart Agriculture System Project:

  • Project Planning and Kickoff
  • Explore the IBM Cloud Platform
  • Connect the IoT Simulator To the Watson IOT Platform
  • Configure the Node-red to Get the Data From the IBM IoT Platform And Open Weather API
  • Building A Web App
  • Configure Your Device to Receive The Data From The Web Application And Control Your Motors

Our Project’s main aim is to help farmers to control their motors from home. He/ She can On and Off his motor by using his mobile phone.
By using Weather API he can know the weather conditions like temperature, humidity, and soil moisture.

Project Background:

  • This Smart Agriculture System Project mainly aims to help the farmers to ease their work.
  • Farmers can get real-time weather conditions by using smart agriculture.
  • Instead of physical devices we create devices in the IBM IoT platform and use them in our project.
  • We connect our device to the IBM node in the NodeRed framework.
  • We need to create a Weather API account to configure the weather API Platform.
  • We then Configure our Node-red to get the weather forecasting data using HTTP requests.

Project Schedule:

  • Project Planning and Kickoff
  • Explore IBM Cloud Platform
  • Connect The IoT Simulator To the Watson IOT Platform
  • Configure The Node-red To Get The Data From IBM IOT Platform And Open Weather API
  • Building A Web App
  • Configure Your Device To Receive The Data From The Web Application And Control Your Motors

Project Requirements:

  • IBM Cloud Account and IBM Watson IOT Platform to create device and sensor
  • Python IDE
  • Node-Red
  • Open weather API Platform

Functional Requirements:

  • Measure Temperature.
  • Gauge Temperature.
  • Gauge Humidity.
  • Gauge Pressure.
  • Weather API.
  • Display the sensor readings using the Watson IOT sensor.
  • Respond to sensor readings and send alerts to the user.

Technical Requirements:

IoT Simulator

Software Requirements:

  • Python
  • Node-Red
  • IBM Watson IOT Platform
  • Open Weather API

Project Deliverables:

A Smart Agriculture System web App for farmers where he can:

• Monitor temperature, humidity, and Soil moisture along with weather forecasting details.
• Control motor for watering the crop through the web app from where he was.

THEORETICAL ANALYSIS

  • Required Software Installation
  • Node-Red

Installation:

  • First, install the Node
  • open command prompt
  • Type ->npm install node-red

To Run the application:

  • open command prompt
  • And then type “node-red”
  • Now open http://localhost:1880/ in the browser

Installation of IBM IOT nodes and Dashboard nodes for Node-Red

  • In order to connect to the IBM Watson IOT platform and create the web UI, these nodes are required
  1. IBM IoT Node
  2. Dash Board Node

IBM Watson IOT Platform

  • Steps To Configure:
  • Create an account in the IBM cloud using your email ID
  • Create IBM Watson Platform in services in your IBM cloud account
  • Launch the IBM Watson lot Platform
  • Create a new device
  • Give credentials like device type, device ID, Token
  • Create API key and store API key and token elsewhere

Python IDE

  • Install python 3 Compiler
  • I Installed PyCharm Community Edition 2020

IoT Simulator

 In our project in the place of sensors, we are going to use a lot sensor simulator which gives random readings to the connected

OpenWeather API

Building Project

Connecting IoT Simulator to IBM Watson IOT Platform

  • Open link Provided in section 4
  • Give the credentials of your device in IBM Watson IoT
  • Click on Connect
  • My credentials given to the simulator are:
  • Organization ID:ka1gns
  • Device Type:nodemcu
  • Device ID:1234five6789
  • Authentication Method:use-token-auth
  • Authentication Token:*********
  • You, Will, receive the simulator data in the cloud
  • You can see the received data in Recent events
  • Data is received in this format (JSON)
  • You can see the received data in cards by creating cards on Boards Tab

4.2    Configuration of Node-Red to collect IBM Cloud Data

  • The Node-Red IBM IoT App is added in Node-Red Work The appropriate device credentials obtained earlier are entered into the node to connect and fetch the device to Node-Red
  • Once it is connected Node-Red receives data from the device
  • Display the data using debug node for

Configuration of Node-Red to collect data from Open weather API

  • The Node-Red also receives data from the OpenWeather API by HTTP GET request. An inject trigger is added to perform HTTP requests for every certain
  • The data we receive from OpenWeather after the request is in JSON format

Configuration of Node-Red to send commands to IBM Cloud

  • By using IBM IoT out Node I used to send data from Node-Red to IBM Watson So, after adding it to flow we need to configure it with the credentials of our Watson device.

Adjusting UserInterface

  • By connecting all the flows shown above
  • We can display our UI by clicking on the dashboard tab in Node-red
  • On the above page, we can display the sensor data and motor
  • On this page, we open weather API data is displayed

Intelligent Customer Help Desk with Smart Document Understanding

INTRODUCTION

Overview

We will be designing an application that leverages multiple Watson Airservices (Discovery, Assistant, Cloud function, and Node Red). By the end of the project, we’ll learn best practices of combining Watson services, and how they can be used to build interactive information retrieval systems with

Discovery + Assistant.

  • Project Requirements: Python, IBM Cloud, IBM Watson
  • Functional Requirements: IBM Cloud
  • Technical Requirements: AI, ML, WATSON AI, PYTHON
  • Software Requirements: Watson assistant, Watson

Scope of Work

  • Create a customer care dialog skill in Watson Assistant
  • Use Smart Document Understanding to build an enhanced Watson Discovery collection
  • Create an IBM Cloud Functions web action that allows Watson Assistant to post queries to Watson Discovery

Proposed solution

For the above problem, we are able to put a virtual agent in the chatbot so it can understand the queries that are posted by customers. The virtual agent should train from some insight records-based company background so it can answer queries supported by the merchandise or associated with the company. In other words, some styles of manual will be accustomed to training the bot using AI. Here I’m using Watson Discovery as a tool for implementing AI and getting trained by the owner’s manual.

THEORETICAL ANALYSIS

Block/Flow Diagram

Hardware / Software Designing

  1. Create IBM Cloud services
  2. Configure Watson Discovery
  3. Create IBM Cloud Functions action
  4. Configure Watson Assistant
  5. Create flow and configure the node
  6. Deploy and run Node-Red app

EXPERIMENTAL INVESTIGATIONS

Create IBM Cloud services

Create the following services:

  • Watson Discovery
  • Watson Assistant
  • IBM cloud function
  • Node-Red

Advantages

  • Companies can deploy chatbots to rectify simple and general human queries.
  • Reduces manpower
  • Cost efficient
  • No need to divert calls to customer agents and customer agents can look at other

Disadvantages:

  • Sometimes chatbots can mislead customers
  • Giving the same answer for different sentiments.
  • Sometimes cannot connect to customer sentiments and intentions

APPLICATIONS

  • It can deploy in popular social media applications like Facebook, slack, and telegram.
  • A chatbot can deploy any website to clarify basic doubts of viewer

CONCLUSION

By doing the above procedure and all we successfully created an Intelligent help desk smart chatbot using Watson assistant, Watson discovery,

Node-RED and cloud functions.

FUTURE SCOPE

We can include Watson studio text-to-speech and speech-to-text services to access the chatbot hands-free. This is one of the future scopes of this project.

Development of Online Shopping Bot using IBM Watson

Introduction

Overview

Online shopping plays a great role in the modern business environment. The best option available for customers in pandemic situations is to use chatbots for online shopping. To support customers in a better way, online shopping bot has opened a door of opportunity and advantage to the firms and customers for having a feel of buying items in a better way. The bot helps to introduce the online shop by listing the items available; it also shows the price of the items and takes orders from the customer. If the customer wishes to see the items, the bot also provides images of the items. This facility ensures the customer sees the products live and gives requests to buy items.

Block Diagram:

Flow Chart Diagram:

Purpose

The online shop bot can help the customer to see the list of items available, images of the images, and the price of the items, and also accepts orders for the items. The purpose of this bot is to save valuable time and money on travel.

  • Literature Survey

In this section, we will discuss the existing solutions available for online shopping and the proposed solution to overcome the limitations.

  • Existing Problems and Solutions

In the past decade, people use the internet as a daily service to access emails, perform online tasks, do shopping, etc. Naturally, people have widely started using the internet at shopper stops too. This showed their willingness to do online shopping. This brings huge responsibility to the shop owners to keep up the buyer’s faith in the particular website. The most important points that affect the customer attitude towards online shopping are customer convenience, collection of information, social contact, and customer diversity.

There are several websites available currently to handle online shopping like Amazon, Flipkart, Big Bazaar, etc. Kotler, (2003) has described the Customer buying method in several sequential steps namely learning, information processing, information searching, evaluating the alternatives, decision making, and post-purchase behavior. When using such websites usability and trust also play a major role and these issues to be handled carefully. With all these facilities available, still we could find some gaps in existing website-based online shopping solutions where the user has limited freedom to communicate or ask doubts regarding items and get a feel of having a discussion with shoppers. This limitation can be overcome by using chatbots for online shopping.

  • Proposed Solution

In recent years, many organizations have shown tremendous interest in developing chatbots for online shopping. These chatbots help customers to handle their queries and to provide information on any kind of items requested. The willingness of the customers to use shopping bots also increased enormously due to the interest in shopping using the internet in pandemic times.

Theoretical Analysis

  • Block diagram
  • Hardware /software requirements
  • Processor: Intel i5
  • Memory: 16GB
  • System Type: 64 Bit Operating system
  • IBM Watson Assistant
  • Node-RED UI Generator

Experimental Investigations

The online shopping bot is developed using IBM Watson. The intents, entities, and context variables are generated, and the JSON file can be downloaded.

Advantages and Disadvantages

  1. An online shopping bot helps to see the list of items available for purchase.
  2. The shopping bot provides the details of the items requested by giving its image and cost.
  3. The shopping bot accepts the mail id to send the order receipt.
  4. The shopping bot accepts the order by asking the item, quantity, and mode of
  5. The shopping bot is interactive.
  6. The shopping bot is simple.
  7. The shopping bot is Usable.
  8. The shopping bot is a user
  9. The shopping bot is available 24/7
  10. The shopping bot is reliable.

Disadvantages

  1. The shopping bot is currently not accepting addresses in the chatbot.
  1. After receiving the receipt, the other interface like email to be used to share the address with the shopper.

Applications

The online shopping bot can be used for advertisement, recommendation, and taking orders from the customer when the customer is living in a chatbot with the shopper.

Conclusion

The online shopping bot is the most useful feature for online shoppers to have a satisfying purchase. With all the possible features embedded in bot it can help the customer to have a successful and satisfactory shopping with less money and time.

Future Scope

  1. The shopping bot can be extended to get reviews
  2. The shopping bot can be added with features like showing offers.
  3. The shopping bot can give recommendations by showing the associated items

Predicting Life Expectancy Using Machine Learning Python Project

Project Name: Predicting Life Expectancy Using Machine Learning

Project scope: The scope of this project is ” Predicting Life Expectancy Using Machine Learning” in this project we are given the task to predict life expectancy, life expectancy is the average time period for which the subject lives.

Project schedule: 

  1. Understanding what to do in the above-given Predicting Life Expectancy Project
  2. Identify and get familiar with the tools needed to complete this project
  3. Writing codes
  4. Collecting Data sets
  5. The time duration is 5 days

 Deliverables: 

  1. Predicting Life Expectancy Using Machine Learning.
  2. Making a user interface too as front-end work and writing code as backend work to make the user interact and calculate the Life

Setting The Development Environment

  1. Creating GitHub account
  2. Creating Slack account
  3. Signing up for cloud services
    1. Node-Red for front end
    2. Watson Studio for coding
  • Machine Learning services

1.  INTRODUCTION

  1.1          Overview:

 This project is based on predicting the life expectancy of a person. It is the statistical average of the number of years a person is expected to live. Factors affecting life expectancy are Country, Mental and Physical Illness, lifestyle, diet, health care services, financial condition, BMI, alcohol consumption, Diseases, etc.

Here in this Predicting Life Expectancy project, our motive is to find the life expectancy of a person after providing details such as the country he is living in is developed or is developing, BMI of the person, Disease history, Income, Population of that country, Expenditure, etc. So here I have used Machine learning and Artificial Intelligence to predict life expectancy. The data used in the training of the model was the data by WHO taken from Kaggle.

There were almost 22 columns stating different factors affecting Life expectancy and 2939 rows comprising data of different persons from different countries. Based on the results we got on Watson Studio some factors which were not affecting the Life expectancy much were removed and the scoring endpoint was obtained after running full code. This scoring endpoint is the URL that helps us to send payload data to a model or function development for analysis (such as to classify the data or to make predictions).

After obtaining the endpoint the next step is to work on Node-red which is the platform, we can use for developing our front-end page that will have a form asking you your details such as year of birth, adult mortality, infant deaths, BMI, etc. rest we’ll discuss in details later on.

Requirements: IBM Cloud, GitHub, Slack, IBM Watson, Node-Red

1.2          Purpose:

 The purpose of this project is to build a model that will predict the Life Expectancy of a person after giving the details of the BMI, Expenditure, Disease history, etc.

2.  Literature Survey

2.1          Proposed Solution:

 The project tries to create a model based on data provided by the World Health Organization (WHO) to evaluate the life expectancy for different countries in years. The data offers data on different person’s Physical health, Mental health, etc from the time frame 2000 to 2015. The data was taken from the website: https://www.kaggle.com/kumarajarshi/life-expectancy- who/data. 

3.  Theoretical Survey

 3.1          Block diagram

 

Block Diagram for Predicting Life Expectancy with Python

3.2          Hardware/Software Designing:

·       GitHub

GitHub is the largest community of developers in the world with millions of people sharing their projects, and ideas for benefiting many people in a very unique way. Any person living in any corner of this world can access this platform for his/her benefit. They can share their problem, their ideas, and solution to some problems. In simple words, it is basically a platform in which anyone can come and share their problems and solutions. It is easy to manage. A team working on the same project can easily monitor the progress and can easily access their work anywhere.

·       Slack

It is a messaging tool that is intended to contact your internal team easily. As it gives you a platform through which we can communicate to our team members easily under one roof. It is not as hectic as sending emails and reading them. It directly comes as a message to you from the group created having your team members. It is great if you are having a team of more than 2 members. Searching for messages becomes an easy, fast medium, searching old messages.

·       IBM Cloud

It is the platform that enables us to use its various features such as Watson Studio which provides a platform where we can write our python code and observe our results in the form of heat maps, graphs, and tables. In this project, we used it and got our scoring endpoint. It is the URL that helps us to send payload data to a model or function development for analysis (such as to classify the data or to make predictions).

·       Node-Red

Node-Red helps us to create a front-end window on which we can get the data from the user such as his Year, BMI, Alcohol intake, etc. and it will then connect to the code written on Watson Studio via the scoring endpoint created after running the python code.

4.  Experimental Investigations

 The graphs of various Factors affecting the prediction of life expectancy are shown in the figure given below:

Curves of life expectancy v/s different factors

Heat map of different factors

Shown above is the heat map of the various factors affecting other various factors some of them have positive values some of them have negative but the thing we have to keep in mind is that we can’t neglect the factors having negative values because they will have the adverse effect which will affect the life expectancy. After some observations, I decided not to include 6 factors that are not affecting life expectancy much and will reduce the calculations and make our model less complex.

5.  Flow Chart

  1. Result

 After filling in all the necessary details asked in the UI form, we got the prediction of life expectancy. The accuracy of our model was 94.41%

Screenshot of the prediction of life expectancy obtained

Advantages and Disadvantages

 Advantages:

  • Easily identifies trends and patterns
  • Wide Applications
  • Handling multi-dimensional and multi-variety data
  • No human intervention is needed (automation)
  • Continuous Improvement

Disadvantages:

  • High error-susceptibility
  • Needs a lot of time to implement
  • Interpreting the results accurately
  • Data set collection is a complex task

Applications 

  1. The form created is easy to understand and is easy to fill by anyone.
  2. It can be used for monitoring health conditions in a particular country
  3. It can be used to get the data about the factor affecting Life expectancy the most in order to work in the direction of obtaining a high life expectancy
  4. It can be used to develop statistics for a country’s development process

Conclusions

This user interface enables any user to predict the life expectancy value of anyone on the basis of the details asked in the form.

Future Scope 

  1. Increase model accuracy
  2. Gives suggestions on how to increase Life Expectancy
  3. Mental health data was missing from the WHO data set which also plays the important role in affecting life expectancy
  4. The scalability and flexibility of the application can be

Energy Management System Python Project

ENERGY MANAGEMENT

  • This is a project on energy management.
  • The project helps to save energy.

Description

  • First, create an IBM cloud account with an e-mail address and password.
  • Then create an IoT platform and node-red platform.
  • We need to write a python code for it because it is a real-time example.
  • Create an MIT app.

First, create an IBM cloud account with an e-mail address and password

Go to google and search for IBM and create an account using your email address and log in to the page. we will get the interface as an IBM dashboard.

CREATING IBM PLATFORM

After creating an IBM account there will be a search option then we can able to see it. Type Internet of things there will be a platform called the internet of thing in that u can able connect the so many services in that internet of things and create a service in that it will ask u to launch it and for creating an internet of things u need some keys as API keys, Device connects and so many things will fill and u can able to see an option as the add device click on new device u can see the device has been created for that u need to keep the API keys, Device information safely in the note pad because it will help u at the need for connecting it.

WRITE PYTHON CODE

we are writing python code because we are going to deal with a real-time example so we need to use python. actually, our project is about how much electricity is causing monthly and its estimated cost
In this, we are writing about fan, light, washing machine, tv, and ac let s take some random values in that because we can’t able take the exact values, and then we need to add all of the to know the charge and then we need to find out the estimation money for it. after completion of this code, u can able to know the code is running.
we need to send it to the IoT platform so we need to install pip by command prompt and then write the code separately so u can able to run and send information to the IoT platform. For sending to the IoT platform we need to give information in a notepad.

python code:

python running code

Information on IoT platform

Creating node-red platform

For creating a node-red platform we need to search in IBM as a node-red app then u can able to see the node app then create it as the local u can create node-red is nothing but we will get the information in the form of a flow chart.
we need to install IoT in, IOT out will be in the node-red app itself and then u can able to use it for better ability and experience we will use the gauge tool and for doing it on the web we will use the HTTP in, HTTP response and then we will use the payload option to print in and I will show u the flow chat I had created and when we run the program u can see the outputs are coming in the debug option.

Create an MIT app.

Search for the MIT app and click on create apps.

First, we should design the app page with alignments given on the left side of MIT. with horizontal alignments drag them to the screen and edit those with suitable markings.
And next insert a text box into the screen. As we have taken energy management the as a fan, ac, light, tv, etc.
Now insert 5 text boxes to the screen and assign the markings. Take two buttons one as checked and another as not checked.

And the two text boxes with one as charge gives the charge of every applicant and another as cost, which gives the summation of all applicants.

This is done and next tap to blocks button. In that create the blocks with URL.

Paste the URL which we have I IBM as:

Now tap on the build option and get the QR code and scan the code you will receive an apk file and install the file and app.

Now the click on the checked button now the values, cost, and charges will be displayed on the screen in their respective blocks.

Covid-19 Testing Management System Python Project

Covid-19 Testing Management System is a small project developed using Python programming. Here are the application features below.

Project features:

• Adding new testing centers
• Search for available testing centers by locality
• Update and delete testing centers
• Show all testing centers by city, state
• Shows the number and type of testing kits available at each center.

Software to be used to develop this application:

• Python
• SQL for creating a database
• Library to interface between Python and SQL
• Either Django (to create a web app) or a GUI library to create the UI (not decided yet)

Implementation

1. Login page (This has a simple login prompt designed using a GUI library that asks for the user id and password.)
2. Menu:
a. Add a testing center
B. Delete testing center
C. Edit the number of testing kits available
D. Search for the testing center by district/city/state
E. Show all testing centers
3. Separate sections for:
a. Adding a new center
b. Deleting center
c. Updating center information
d. Searching for center
e. Nationwide data

Covid-19 Testing Management System can be developed by using PHP & MySQL Server with different functionalities below.

User Characteristics

The Covid 19 Testing Management system has 2 types of users they are Admin and user(patient)

General Constraints

The tools and technologies that are used to develop this project are:
The language used in this project is PHP5.6 and PHP7.x.
The database used in this project is MySQL 5.x.
The web browsers that are used in this project are Mozilla, Google Chrome, IE8, and OPERA.

Operational Scenarios

Scenario A:
How your application starts
Our application covid19 testing management system starts by login into the application, if the user is a new user, the user needs to register by providing the needed credentials.
New users need to provide testing information. A registered user needs to provide test information.

Scenario B:
Usage Scenarios like Customer Check-out
When a customer visits our application, he/she needs to log in by providing their credentials according to their role i.e., admin and user.
If the user is new, he needs to register if he is an old user, he can directly enter the login details and login directly.

Scenario C:
Database

The data that we are going to store in the database
1) Admin Login
Username
Password

2) New admin login
Admin name
Username
Password

3) User login
Username
Password

4) New user login
Name
Phone number
Username
Password

5) User Registration
Name
Current address
Gender
D.O.B
E-mail id
Phone number
Age
Aadhar card number

Classification of American Sign Language using online RESTful application

  • INTRODUCTION

This document report provides the desired layout to develop an online application service that accepts Human Skeletal key points of a sign video and returns the label of the sign in a JSON response. The document contains information about the extraction of key points from the videos using Tensor Flow’s Pose Net library and four different deep learning models that can classify American Sign Languages into six different signs. i.e {buy, fun, hope, really, communicate, mother}. Moreover, it also contains information about hosting services using flask API on ‘PythonAnywhere’ and steps involving handling HTTP requests coming from different users.

  • TECHNICAL APPROACH

Firstly, we have accumulated all the raw video data sets which have been recorded as a part of Assignment-1 and extracted frames of the particular timeline. Then, we used Tensor Flow’s Pose Net library in order to extract key points from the images, which are considerably used as training data for models. We have tried three different approaches to preprocess data and picked the one which gives the best accuracy for the trained models.

Approach-1: Scaled down raw data using the Universal Normalization technique and extracted a few features like- Standard Deviation, Moving Mean of Window size 5, Zero Crossing Rate, Dynamic Time Warping distance, and built feature matrix. Then we applied PCA on the feature matrix and using K-fold Cross-validation we trained four deep learning models named Convolutional Neural Network, K nearest neighbor, Support Vector Machine, and Random Forest. The average accuracy of the given models lay between 60-65%.

Approach- 2: As a part of the second approach, we expelled some features by observing the movement of each body part in videos for different signs and made a feature matrix of only important features. Then we apply Standard Scaler and Min Max Scaler in order to normalize data and trained our models using the first approach.

Somehow, we were able to increase the average accuracy of the models by 10%.

Approach -3: We have observed in the second approach that, our model is only considering the static coordinates of each body part, so we subtracted each coordinate of different body parts from the static body parts and processed the data in the same manner. So, by doing this approach we got our highest average accuracy which lies between 85% to 90%. 

  • INITIAL FEATURE EXTRACTION 
  1. Zero Crossing Rate
  2. Moving Average Window
  • Standard Deviation
  1. Dynamic Time Warping Distance.

Zero Crossing Rate: The zero crossing rate is the rate of sign- changes along with a signal, i.e the rate at which the signal changes from positive to zero to negative or vice versa. Zero Crossing Rate can be used as a primitive pitch detection algorithm for signal processing.

Moving Average Window: Moving Average is optimal for reducing random noise while retaining a sharp step response. This makes it the premier filter for the time domain encoded signals

Standard Deviation: The standard deviation is a measure of how far the signal fluctuates from the mean. It also depicts how data disperse near the mean of particular data series.

Dynamic Time Warping Distance: DTW measures the similarity between two temporal series data. Any linear sequence data can be analyzed with DTW, it aims at aligning two sequences of feature vectors by warping the time axis iteratively until an optimal match between the two sequences is found.

Feature Engineering:

We have expelled a few features by observing the movement of each body part for different signs and made a feature matrix with only important features. Below is the list of features that we considered for training models.

[“nose_x”, “nose_y”, “leftShoulder_x”, “leftShoulder_y”, “rightShoulder_x”, “rightShoulder_y”,”leftElbow_x”, “leftElbow_y”, “rightElbow_x”, “rightElbow_y”, “leftWrist_x”, “leftWrist_y”,      “rightWrist_x”, “rightWrist_y”]

Here, we observed that the coordinates value of each body part shows a static position for a given time, so we have subtracted each body part’s coordinates value from the corresponding static body part’s coordinates. Here, we have considered “nose”  as a static body part and subtracted each body part with corresponding X and Y coordinates.

The above-mentioned approach would become simpler for the models to understand the movement of each body part, as we have a relative position for each body part, the model can easily predict certain gestures by examining the positive or negative sides of coordinates.

  • MODELS USED:
  1. K nearest neighbor
  2. Convolution Neural Network
  • Support Vector Machine
  1. Random Forest

K Nearest Neighbor: The K Nearest Neighbor classifier is one of the most simple machine learning algorithms that simply relies on the distance feature vectors. It classifies unknown data by finding the most common classes among k nearest examples. The majority vote of the class label is assigned to unknown data. As KNN is a lazy learning algorithm, it works more efficiently when our dataset has been distributed in multi classes.

Support Vector Machine: The core idea of Support Vector Machine is to find a hyperplane that separates two sets of objects having different classes. It uses a technique called kernel trick to transform data and based on this transformation it finds an optimal boundary. It is considered one of the most robust and accurate algorithms among other classifiers.

Random Forest: Random Forest is an ensemble classifier; it takes multiple individual models and combined them into a more powerful aggregate model. So, let’s say we have different individual models, then there might be the case that they work efficiently because some part of the data set would be overfitted to the model. So by combining them, we can reduce the chances of error. Random forest built upon by aggregating n possible decision trees which might be generated by randomly picking data set row as root. So, as the dataset would be increasing, the possibilities of generating random decision trees would also increase and aggregating different decision tree models lead to an increase in the efficiency of the aggregated model, too.

REFERENCES:

Drowsiness Detection of a Cab Driver Python Project

  • In the present situation world, Accidents are part of life. On an average of 1214 road accidents in India only. One of the main reasons for road accidents is being drowsy during driving.
  • The application Drowsiness Detection of a Cab Driver helps the manager of the company get a notification that the driver is drowsy and an alarm rings in the car to alert the driver.

USAGE:

>    python detect_drowsiness.py –shape-predictor shape_predictor_68_face_landmarks.dat –alarm alarm.wav

Use the above command to run the application

This Drowsiness Detection of A Cab Driver application needs the following modules: –

MODULE NAME:  scipy

COMMAND USED FOR INSTALLING: conda install -c anaconda scipy

MODULE NAME: imutils

COMMAND USED FOR INSTALLING: conda install -c anaconda imutils

MODULE NAME: threading

COMMAND USED FOR INSTALLING: conda install -c anaconda threading

MODULE NAME: numpy

COMMAND USED FOR INSTALLING: conda install -c anaconda numpy

MODULE NAME: playsound

COMMAND USED FOR INSTALLING: conda install -c anaconda playsound

MODULE NAME: argparse

COMMAND USED FOR INSTALLING: conda install -c anaconda argparse

MODULE NAME: time

COMMAND USED FOR INSTALLING: conda install -c anaconda time

MODULE NAME: cv2

COMMAND USED FOR INSTALLING: conda install -c anaconda cv2

MODULE NAME: dlib

COMMAND USED FOR INSTALLING: It should be installed in the virtual environment the command is: conda install -c anaconda dlib

IN THE ABOVE-MENTIONED MODULES MOST OF THEM WILL BE GIVEN DEFAULT

TO CHECK THE MODULES THE COMMAND IS conda list

Download the complete project source code