86

Download Download PDF - International Journal of Intelligent

Embed Size (px)

Citation preview

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 248–250 | 248

A Review of Smart Parking System based on Internet of Things

Harkiran Kaur 1, Jyoteesh Malhotra2

Accepted : 08/11/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: The Internet of Things expands the wireless paradigm to the edge of the network, which enables to develop a different kind of

applications or services. By using IoT the application becomes more reliable, flexible or portable. It has a large amount of nodes on the

various geographic positions, to increase the awareness of location or reduce the latency. There are numerous IoT applications like smart

grid, smart hospital, smart industry, smart traffic management etc. In this paper, we describe the smart traffic management and parking

system using the internet to control the chaos and also discuss the various researches on this concept.

Keywords: Bluetooth, Cloud Computing, Information Communication Technology (ICT), Internet of Things (IoT), Wireless Sensor

Network (WSN)

1. Introduction

The use of IoT in the practical world become more relevant in the

recent year by the rapid growth of technologies like mobile,

embedded system, wireless system, cloud computing etc. Internet

of Things helps to connect the living beings and the non-living

things, which generates the communication between the virtual

and non-virtual data. So, there are a large amount of data

generated by the various number of nodes connected via the

internet. After that, the relevant information should be converted

from the pool of data. The WSN plays a very imperative role in

the development of IoT. The concept of the smart city is the best

suitable example for the integration of IoT and WSN with ICT.

Smart City provides an advanced step for groundbreaking and

cooperative amenities for all the social events such as healthcare,

transportation, business etc. It uses IoT in an effective and secure

manner to access the IP address of the object, and generate the

communication link between the object and the system with the

help of internet.

2. Related Work

IoT has a vital role in the traffic management in smart cities.

Nowadays, many techniques like wireless sensor network, radio

frequency, digital image processing based smart transportation

system are introduced for traffic management. There are used to

control the traffic in the more efficient way. In 2012, J. Yang uses

GPS (Global Positioning system) and Android based application

to provide the information related to the parking space. It has a

better interface but it does not have the reservation feature [1].

M. Patil implements the WSN and RFID for the smart parking in

2013 [2]. In this system, the Inter-Integrated Circuit protocol is

used according to that the car will guide the driver to find the

path for parking by using RFID and WSN. This system is time-

consuming and expensive which are the major drawbacks. H.

Singh introduced the automated parking system with the help of

Bluetooth access. In this system, the Bluetooth is used for the

communication or network. This system is based on the rack and

pinion mechanism for the linear motion. So the main drawback of

this system is the whole parking is designed according to the rack

and pinion mechanism which is quite expensive and time-

consuming [3].

According to Hilal Al-Khaurasi, the intelligent parking system

was introduced which is based on image processing. In this paper,

the concept of image processing is used. The sensor camera is

used to capture the images which help to provide the results that

are slots for parking are free or not. But, in term of visibility, the

weather conditions like rain, fog, snow etc. may affect the system

[4]. Another system is made by A.Sayeeraman in 2012. In this

system, ZigBee is used to check the vacancy in the parking area.

In this, the security feature is added i.e. exist password, without

entering the password user cannot leave the parking. The gate

will open when the user put the right exit password. The exit

password is sent to the user at the entry time via SMS. The

problem of this system is network congestion due to which user

unable to receive the SMS. However, the use of SMS, GSM or

ZigBee makes the system more expensive [5]. The other authors

develop the parking reservation system by using the SMS. In this

paper, the microcontroller is used which makes the system

portable but the cost of implementation is very high because of

the use of the microcontroller. The major problem is the system

will crash if the workload on the microcontroller will

increase[6].This [7] paper introduced the use of RFID, WSN,

object ad-hoc network and information systems in which the

traffic objects can be automatically tracked and controlled over a

network. The concept of RFID is used as a new technology to

reduce the installation cost and time [8].According to J. Sherly

IoT-based transportation system is developed according to smart

city view. The main aim of this paper is to develop the advanced

and powerful communication between the technologies for the

administration of the city and the citizens[9]. The raspberry pi 2

is used for the smart high defined picture. There are some

ultrasonic sensors are used to detect the vehicle intensity and

sends the signal to RASPBERRY PI 2 which controls the traffic

problem [10].

Y K. Zang [11] designed an efficient decentralized framework for

reducing the fuel consumption at any time of expedition or

stoppage. This algorithm provides a solution for congestion

________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1Department of Computer Science and Engineering, GNDU RC

Jalandhar, Punjab, India 2Department of Electrical and Computer Engineering, GNDU RC

Jalandhar, Punjab, India

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 248–250 | 249

Table 1. Current State-of-the-art

Sr.

No. Paper Name

Year

Description Merits Demerits

1. Smart Parking Services based on

Wireless Sensor Network[1]. 2012

In this paper, the android application is used. It provides the better

interface but it does not have reservation feature.

Usage of Android platform provides

better interface

Easy to use.

This system has not reservation feature.

2. Wireless Sensor Network and RFID for

the Smart Parking System[2]. 2013

In this paper, the RFID and the wireless sensors are used to make the

Parking, according to this, it guide the driver to find the spot but this

system is time-consuming and expensive.

It provides the slot information as

well as guides the driver.

It is compatible in existing parking.

High cost

More time required for implementation.

3. Automated Parking System with

Bluetooth access[3]. 2014

In this system, the Bluetooth is used for the communication or network.

This system is based on the rack and pinion mechanism for the linear

motion. So the main drawback of this system is the whole parking is

designed according to the rack and pinion mechanism which is quite

expensive and time-consuming.

Bluetooth is used for registration or

identification.

Detect the unique registration number

if a new vehicle is parked.

Cannot be used in the existing system.

Whole parking is to be designed.

Costly.

Time-consuming.

4. Intelligent Parking Management System

based on Image Processing[4]. 2014

In this paper, the concept of image processing is used. The sensor camera

is used to capture the images which help to provide the results that are slots

for parking are free or not. But, in term of visibility, the weather conditions

like rain, fog, snow etc. may affect the system.

Easily detect the presence of vehicles.

The camera is used as a sensor.

Not compatible with some weather

conditions such as rain, fog etc.

Fixed position of the system.

GPS is not provided.

5.

Zigbee and GSM based Secure Vehicle

parking management and reservation

system[5].

2012

In this system, ZigBee is used to check the vacancy in the parking area. In

this, the security feature is added i.e. exist password, without entering the

password user cannot leave the parking. The gate will open when the user

put the right exit password.

Secure

Based on password system.

Use of GSM

Use of SMS

Expensive.

The problem of this system is network

congestion due to which user unable to

receive the SMS.

6. Smart Parking Reservation System using

short message services (SMS)[6]. 2009

In this paper, the microcontroller is used which makes the system portable

but the cost of implementation is very high because of the use of the

microcontroller. The major problem is the system will crash if the

workload on the microcontroller will increase.

More

Secure.

Ease of usage.

The cost of implementation is high.

GSM feature create bottlenecks

Overload of microcontroller leads to

system crash.

7.

Intelligent Traffic Information System

Based on Integration of Internet of

Things and Agent Technology[7].

2015

This paper introduced the use of RFID, WSN, object ad-hoc network and

information systems in which the traffic objects can be automatically

tracked and controlled over a network.

Provides tracking feature.

Use of WSN made this system more

reliable.

Networking congestion.

8. Smart Traffic Management System[8]. 2013 The concept of RFID is used as a new technology to reduce the installation

cost and time. Reduce the installation time and cost. GPS is not used.

9. Internet of Things Based Smart

Transportation System[9]. 2015

In this paper, the IoT-based transportation system is developed according

to smart city view. The main aim of this paper is to develop the advanced

and powerful communication between the technologies for the

administration of the city and the citizens.

Provide instant results.

Use for real-time traffic.

The problem in data storage.

Use of image processing system.

10. Innovative Technology for Smart Roads

by using IoT devices[10]. 2016

In this paper, the raspberry pi 2 is used for the smart high defined picture.

There are some ultrasonic sensors are used to detect the vehicle intensity

and sends the signal to RASPBERRY PI 2 which controls the traffic

problem.

Automatically detect the vehicles.

Controls traffic related problems. Expensive.

11.

Optimal control and coordination of

connected and automated vehicles at

urban traffic intersections[11].

2016

An efficient decentralized framework for reducing the fuel consumption at

any time of expedition or stoppage. This algorithm provides solution for

congestion control and helps to reduce the traveling time

Efficient framework.

Solve congestion problem.

Reduce traveling time.

Large time for implementation.

High cost.

12.

Assessing the mobility and

environmental benefits of reservation-

based intelligent intersections using an

integrated simulator[12].

2012

In this designed intersection control on reservation based system to take

more advantage of extraordinary connectivity that provides dynamism to

connected transports

Online reservation.

Dynamism connectivity of transports.

The cost of reservation is high.

Network congestion.

13.

Analysis of reservation algorithms for

cooperative planning at

intersections[13].

2010

This represents the designed the fully automated car that has the major

differences compared to nowadays. This paper also describes the

improvement of the reservation algorithm.

Based on reservation algorithm Not practically implemented.

14.

Analysis and modeled design of one

state-driven autonomous passing-

through algorithm for driverless vehicles

at intersections[14].

2013

In this paper, proposed the centralized scheduling algorithm i.e.

reservation-oriented. This ensures the high request to be answered as per

the preference. This algorithm is simulated on high priority vehicles

Simulated on high priority vehicles.

Based on centralized scheduling

algorithm

Complex

The difficulty of understanding.

15. Cloud-based Intelligent Transport

System[15]. 2015

This paper represent the multilayered vehicular data cloud platform which

was supported by IoT technologies and cloud computing Support multiple technologies.

The problem in data storage.

Hard to manage a large amount of data.

16.

Visible light communication:

Application, architecture,

standardization and research

challenges[16].

2016

In this paper, the unique computer code is designed that can communicate

with various devices within the road. In this technological era, visible light

communication is also used for vehicle to vehicle communication.

Strong communication between

machines to the machine.

Consume a lot of time for

implementation.

17. Enabling Reliable and Secure IOT-

based Smart City Application[17]. 2014

According to Elias Z. Tragos in this paper the IoT is used as the real-time

traffic monitor to reduce the overall traffic from the downtown area.

Monitor real-time traffic.

Helps to reduce the overall traffic. -

18. Smart City Architecture and its

Application based on IoT[18]. 2015

This paper mainly based on customized services for communication with

the help of Wi-Fi, WI-Max, Zig-Bee, Satellite Communication etc. to

convert the city into the smart city. For instance, we combine the health

and transportation domain by this sensor can easily measure the all health

related parameter of the driver while driving like pulse rate, blood pressure

etc. and provide the real-time health condition to the driver. This helps to

create the safer environment.

User-friendly.

Ease to communicate.

Measure all health related parameters.

Safer environment.

Lots of sensors are used.

Required more time as compared to

another system.

19. Automatic Smart Parking System using

Internet of Things(IoT)[19]. 2015

In this, the system designs a simple smart car parking system which is cost

effective and also helps to decrease the carbon dioxide that makes this

system eco-friendly.

Cost effective

Eco-friendly.

Not solve the congestion related

problems.

20.

Automatic Parking Management System

and Parking fee collection based on

number plate recognition[20].

2012 The author utilizes wide-point camera as a sensor which records or find

free space. These records are used for proper utilization of space. Proper utilization of space.

Use of camera makes the system

unreliable.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 248–250 | 250

control and helps to reduce the traveling time. S. Huang[12]

designed intersection control on the reservation based system to

take more advantage of extraordinary connectivity that provides

dynamism to connected transports. A. Fortella[13] represents the

designed fully automated car that has the major differences

compared to nowadays. This paper also describes the

improvement of the reservation algorithm. The K.Zang [14]

proposed the centralized scheduling algorithm i.e. reservation-

oriented. This ensures the high request to be answered as per the

preference. This algorithm is simulated on high priority vehicles.

K. Ashokkumar [15] represent the multilayered vehicular data

cloud platform which was supported by IoT technologies and

cloud computing. In this paper, the unique computer code is

designed that can communicate with various devices within the

road. In this technological era, visible light communication is also

used for vehicle to vehicle communication [16].According to

Elias Z. Tragos [17] IoT is used as the real-time traffic monitor to

reduce the overall traffic from the downtown area. [18] This

paper mainly based on customized services for communication

with the help of Wi-Fi, WI-Max, Zig-Bee, Satellite

Communication etc. to convert the city into a smart city. For

instance, we combine the health and transportation domain by

this sensor can easily measure the all health related parameter of a

driver while driving like pulse rate, blood pressure etc. and

provide the real-time health condition to the driver. This helps to

create the safer environment. Basavaraju [19] design a simple

smart car parking system which is cost effective and also helps to

decrease the carbon dioxide that makes this system eco-friendly.

The author utilizes wide-point camera as a sensor which records

or find free space and these records are used for proper utilization

of space for the smart parking [20].

3. Current State-of-the-art

Based on literature survey done the following papers have been

abstracted to find out the key merits and demerits.

4. Open Issues and Challenges

Nowadays, there are various challenges which are faced by

today’s driver as well as whole transportation system on daily

basis likewise traffic congestion, parking system etc. To improve

this transportation system, the following issues and factors are

considered:

Be able to sense vehicle accurately.

Provide congestion free road.

Manage a large amount of data.

Enable to make smart decisions using real-time data.

Reduce frustration and improve time-saving.

Alleviate the problem related to pollution emission and fuel

consumption.

5. Conclusion

Overall, in this paper, there are various types of methods or

techniques are discussed. This review provides the gather

information about the traffic management or smart parking

methods which are used for the smart cities. The wide availability

of sensors or wireless devices made the system more reliable or

flexible and enables effective development of the IoT-based

application. IoT provides enormous kind of solution for the traffic

controlling and provide the vehicle to vehicle communication for

the optimal result. This paper helps to understand all various

methods of traffic management under the concept of the smart

city.

References

[1] Yang Jihoon Jorge Portilla Teresa Riesgo, “Smart parking service

based on wireless sensors Network,” IEEE, 2012.

[2] ManjushaPatil Vasant N. Bhonge, “Wireless Sensor Network and

RFID for Smart Parking System,” Int. J. Emerg. Technol. Adv.

Eng. Website www.ijetae.com, vol. 3, no. 4, 2013.

[3] Harmeet Singh ChetanAnand Vinay Kumar Ankit Sharma,

“Automated Parking System With Bluetooth Access,” Int. J. Eng.

Comput. Sci. ISSN2319-7242, vol. 3, no. 5, pp. 5773–5775.

[4] H. A.-K. I. Al-Bahadly, “Intelligent Parking Management System

Based on Image Processing,” J. Eng. Technol., vol. 2, pp. 55–67,

2014.

[5] AshwinSayeeraman P.S.Ramesh, “ZigBee and GSM based secure

vehicle parking management and reservation system,” J. Theor.

Appl. Inf. Technol., vol. 37, 2012.

[6] HanitaDaud Noor HazrinHanyMohamadHanif Mohd Hafiz

Badiozaman, “Smart parking reservation system using short

message services (SMS),” IEEE, 2009.

[7] H. O. Al-Sakran, “Intelligent Traffic Information System Based on

Integration of Internet of Things and Agent Technology,” Int. J.

Adv. computer Sci. Appl., vol. 6, no. 2, 2015.

[8] N. Lanke, “Smart Traffic Management System,” Int. J. Comput.

Appl., vol. 75, no. 7.

[9] J. Sherly D. Sonasundareswari, “Internet of Thing Based Smart

Transport System,” Int. J. Eng. Technol., vol. 75, no. 7, 2013.

[10] Sheela.S, “Innovative Technology for Smart Roads by Using IOT

Devices,” Int. J. Innov. Res. Sci. Eng. Technol., vol. 5, no. 10.

[11] Y. J. Zhang or A. A. Malikopoulos and C. G. Cassandras, “Optimal

control and coordination of connected and automated vehicles at

urban traffic intersections,” in Proc. Amer. Control

Conference[online], 2016.

[12] S. Huang or A. W. Sadek and Y. Zhao, “Assessing the mobility

and environmental benefits of reservation-based intelligent

intersections using an integrated simulator,” IEEE Trans. Intell.

Transp. Syst., vol. 13, 2012.

[13] A. d. La Fortelle, “Analysis of reservation algorithms for

cooperative planning at intersections,” in Proceding. 13th

International IEEE Conference Intelligent Transport System, 2010.

[14] K. Zhang A. de La Fortelle D. Zhang and X. Wu, “Analysis and

modeled design of one state-driven autonomous passing-through

algorithm for driverless vehicles at intersections,” in Proc. IEEE

16th Int. Conf. Comput.Sci. Eng, 2013.

[15] K. Ashokkumar, “CLOUD BASED INTELLIGENT

TRANSPORT SYSTEM,” Sci. Direct, vol. 50, pp. 58–63, 2015.

[16] L. U. Khan, “Visible light communication: Application,

architecture, standardization, and research challenges.,” Digit.

Commun. Networks, Elsevier, pp. 1–10, 2016.

[17] E. Z.Tragos, “Enabling Reliable and Secure IOT- based Smart City

Application,” IEEE, pp. 111–116, 2014.

[18] G. Aditya, S. Bryan, P. Gerard, and M. Sally, “Smart City

Architecture and its Application based on IoT,” Sci. Direct,,

Elsevier, vol. 52, pp. 1089–1094, 2015.

[19] S. R. Basavaraju, “Automatic Smart Parking System using Internet

of Things(IoT),” Int. J. Sci. Res. Publ., vol. 5, no. 12, pp. 629–632,

2015.

[20] R. M.M, A.Musa, AtaurRahman, N.Farahana, and A.Farahana,

“Automatic Parking Management System and Parking fee

collection based on number plate recognition,” Int. Mach. Learn.

Comput., vol. 2, 2012.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 251

Text line Segmentation in Compressed Representation of Handwritten

Document using Tunneling Algorithm

Amarnath R.*1, Nagabhushan P. 1

Accepted : 01/10/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: In this research work, we perform text line segmentation directly in compressed representation of an unconstraint handwritten

document image using tunneling algorithm. In this relation, we make use of text line terminal point which is the current state-of-the-art

that enables text line segmentation. The terminal points spotted along both margins (left and right) of a document image for every text

line are considered as source and target respectively. The effort in spotting the terminal positions is performed directly in the compressed

domain. The tunneling algorithm uses a single agent (or robot) to identify the coordinate positions in the compressed representation to

perform text-line segmentation of the document. The agent starts at a source point and progressively tunnels a path routing in between

two adjacent text lines and reaches the probable target. The agent’s navigation path from source to the target bypassing obstacles, if any,

results in segregating the two adjacent text lines. However, the target point would be known only when the agent reaches destination; this

is applicable for all source points and henceforth we could analyze the correspondence between source and target nodes. In compressed

representation of a document image, the continuous pixel values in a spatial domain are available in the form of batches known as white-

runs (background) and black-runs (foreground). These batches are considered as features of a document image represented in a Grid.

Performing text-line segmentation using these features makes the system inexpensive when compared to spatial domain processing.

Artificial Intelligence in Expert systems, dynamic programming and greedy strategies are employed for every search space while

tunneling. An exhaustive experimentation is carried out on various benchmark datasets including ICDAR13 and the performances are

reported.

Keywords: Compressed Representation, Handwritten Document Image, Text-Line Terminal Point, Text-Line Segmentation, Search

Space, Grid.

1. Introduction

Technological advances of storage and transfer have made it

possible to maintain many documents in the digital format. It is

also necessary to preserve these documents in digital image

format only, particularly in case of handwritten documents for

verification and authentication purposes [1, 2, 3, 4, 5].

Maintaining these document images in the digital form would

require huge storage space and network bandwidth. Therefore, an

efficient compressed representation would be an effective

solution to the storage and transfer issues [6, 7, 8] particularly

arising from document image. Generally, the document image

compression format follows the guidelines of CCITT Standards

[9, 10] which is a part of the ITU (International Telegraph

Union). These standards were specifically targeted towards

document images stored in digital libraries. On the other hand, if

document images, audios and videos are to be archived and

communicated in the compressed form itself, then it would be

considered as a third advantage in addition to storage and

transmission.

The digital libraries with document images in their compressed

formats could imply a solution to a big data problem arising from

the document images, particularly about storage and transmission

[11]. The compressed image file formats such as TIFF, JPEG,

and PNG, strictly follow CCITT standards [7, 8, 9, 10, 11, 12].

For any digital document analysis (DDA), the image in its

compressed format must undergo the decompression stage.

However, performing operations on decompressed documents

would unwarrantedly suppress the advantages of both the time

and buffer space [6, 13, 14, 15]. If DDA could be achieved in the

compressed version of the document image without

decompression, then the document image compression could be

viewed as an effective solution to the immense data problems

arising from document images [13].

The idea of operating and analysing directly in the compressed

version of the document image without decompression is known

as Compressed Domain Processing (CDP) [6, 13, 14, 15). A

recent literature [12] on CDP shows the strategies to perform

document image analysis in its compressed representation.

However, the model is subjected to the printed documents only.

The challenging job is to perform DDA in compressed

representation of handwritten document image. Performing DDA

on uncompressed handwritten document could be a difficult task

because of oscillatory variations, inclined orientation and

frequent touching of text lines while scribing on un-ruled paper

[13]. An initial attempt [13] to spot the separator points at every

line terminals in compressed unconstrained handwritten

document images using run-length features (Run Length

Encoding or RLE) is the state-of-the-art technology. This

motivates to carryout text-line segmentation in compressed

document image. In this research, we have attempted to segment

the text lines in the compressed representation with the help of

these spotted terminal points. In this context, a research [16]

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Department of Studies in Computer Science, University of Mysore,

Karnataka, India

* Corresponding Author Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 252

shows a method of extracting run-length features from the

compressed image format supported by CCITT group 3 and

CCITT group 4 standards. These protocols use run-length

encoding, which is widely accepted for binary document image

compression. Our paper aims to segment the text lines of the

document image using the run-length data which is represented in

a grid. The detailed explanation about the RLE format of the

document image is provided in Section 3.

The current state-of-the-art uses a single column (first column) of

the grid to find the separator points at every line terminal along

left margin of the document image. But the last column of RLE

does not infer the depth of right margin of the document for

spotting separator points. To work on the right margin of the

document image, the final non-zero entry of every row in RLE

data was observed for creating a virtual column and thus

separator points are spotted. Now, these identified separator

points at both the margins are considered as source (separator

points along left margin) and target (separator points along right

margin) nodes. In this paper, we make use of single agent (or

robot) for text-line segmentation. The agent’s job is to tunnel or

trace the path by navigating between the two-adjacent text-lines

starting from a source point at one end of a document and

reaching the destination point present at the other end of the

document. This process is applied to all the source nodes

resulting in segmentation of all text-lines. Some interesting

observations inculcated includes analysing the correspondence

between the source nodes and terminal nodes. We also addressed

some of the issues related to wrong segmentation that is when

two or more paths converge to one destination traced by the

agent. Corrective measurements are taken to tackle this issue by

understanding the correlation between two adjacent paths.

Our approach identifies the text line positions operating directly

in the RLE representation of the document images which results

in text-line segmentation. The rest of the paper is organized as

follows. A survey of related research is presented in section 2. In

section 3, we have explained the RLE representation of a

document image. Further, we provide the terminologies and

corollaries used in this paper. Section 4 describes the algorithmic

modelling along with comparative time complexity with respect

to both RLE and uncompressed document versions. Further,

experimental results conducted on benchmark handwritten

datasets such as ICDAR 13 and other databases [17] are

described in Section 5. Section 6 summarizes the research work

with future avenues.

2. Related Works

In the recent past, we could trace few related works on CDP, but

restricted to printed document images. A literature [15] on CDP

provides a detailed study on document image analysis techniques

from the perspectives of image processing, image compression

and compressed domain processing. This enabled various

operations carried out in the field of skew detection / correction,

document matching, document archival.

Recent article [18], demonstrates straight line detection in RLE

representation of the handwritten document image. Incremental

learning-based text-line segmentation in compressed handwritten

document images are reported in literature [19]. Further, there

was a technical article [13] that discusses about performing direct

operations on the compressed representation of handwritten

document. The effort is to spot the separator points at every text

lines in both margins of the document image enabling text line

segmentation. For better understanding, the identified line

terminal points applied over an uncompressed document image is

shown in Figure 1.

Fig. 1. Line Separators at both borders of a document in uncompressed

documents (Reference: A portion of ICDAR13 test image - 320.tif)

One of the significant models [20] uses seam carving approach to

segment the text-line that works on historical documents of

uncompressed images. Another recent study [21] discusses

various path finding algorithms which are a class of heuristic

algorithms based on Artificial Intelligent. These techniques are

domain specific that works on uncompressed document images.

However, the ideas of path finding approaches are inculcated in

our proposed model, that is to operate on compressed version for

decision-making in every search space. Further, an effort was

made to carry out the text-line segmentation directly in

compressed handwritten document images to avoid

decompression [13] which is the main objective of this paper.

3. Compressed Image Representation and Corollaries

The Modified Huffman (MH) [7, 8] is the most common

encoding technique following CCITT compression standards

which is supported by all the facsimiles. The improved

compression versions are Modified Read (MR) [7, 8, 9, 10] and

Modified Modified Read (MMR) [7, 8, 9, 10]. A comparative

study on encoding / decoding techniques of these compression

standards are tabulated (Table 1). MH uses RLE for an efficient

coding, whereas MR and MMR exploit the correlation between

successive lines of a document image.

Table 1. Compression standards

Characteristics MH MR MMR

CCITT Standard T.4 T.4 T.6

Dimension 1-D 2-D 2-D (Extended)

Algorithm Huffman and

RLE

Similarity

between two successive lines

Extended MR

Figures 2, 3 and 4 show a sample binary image, its corresponding

hexadecimal image viewer and the RLE representation of the

image respectively. The RLE representation of an image is

defined as Grid (G) in this paper.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 253

Fig. 2. A sample image of size 18X10 (WidthXHeight)

Fig. 3. Hexadecimal visualization of the image from Fig. 2

Fig. 4. RLE data inferred as Grid of the image from Fig 2

The odd and even columns of the RLE data represent the count of

white-runs and the black-runs of an uncompressed document

image. The number of continuous pixel value, say ‘0’

(background), is called as white-runs [22, 23, 24, 25, 26].

Similarly, the number of continuous pixel value, say ‘1’

(foreground) is called as black-runs [22, 23, 24, 25, 26]. If a row

in an image starts directly with foreground, then the value in the

first column of the corresponding row in RLE has an entry zero

(‘0’). This infers that odd columns and even columns in the grid

contain count of white-runs and black-runs respectively. The

white-runs represent background whereas the black-runs

represent foreground as text-content.

Some of the corollaries used in this article are listed here:

a. Each separator points spotted in the first column of the Grid

( ) is considered as Source Node ( ).

b. Each separator points spotted in the virtual column is

considered as Target Node ( ). Final non-zero entry in

every row of G makes this virtual column.

c. A path ( ) is defined as segmentation of adjacent text lines;

an agent (or robot) navigates along blank space (white-

runs) available between two adjacent text-lines from to

with reference to .

d. Edges ( ) are defined as sequences of white-runs that exist or

tunneled between two adjacent text-lines. These sequence

of formulates .

e. An obstacle ( ) appears because of touching characters

between two adjacent text-lines, which is a hurdle while

tracing from to . This may also occur due to ascenders

and descenders of characters, which appears larger in length

when compared to a given threshold ( ).

f. t is the length covered by an agent tracing both the direction

vertically up and down ( ) from the current

position , where arguments and refer to the

position in along (columns) and (rows) axis

respectively; this search space is induced whenever there is

an . is chosen empirically based on experiments

conducted on the datasets.

g. For every , there is a . The relationship between and is

one-to-one function defined as and strictly

follows the ascending order. The correspondence of and

is not defined in the literature [13]. However, in this

paper, we could find the probable correspondence between

and with help of an agent (or robot) and algorithmic

strategies.

h. When relationship between and is onto function or

crossover, then it is observed as wrong segmentation. Even

though and hold a relation one-to-one and there exist

crossovers, then it is considered as wrong segmentation.

i. The distance ( ) calculated between and denotes the

shortest covering longest distance (weights), which is

defined by a function, . depends on the number of

edges ( ) used in by an agent to cover between and

.

j. The terminal points and does not necessarily possess

larger white-runs because, they are heuristically chosen as

mid-points of the bands corresponding to the two-adjacent

text-lines along axis (columns) of . Therefore, may

not be an actual start point, in the given situation, whereas

the weight of the is longer than . So, the actual start

point would be and it would be within the search space of

from .

k. The distance between two intermediate hubs such as

and must be equal or greater than a given threshold ( ).

is calculated by finding a maximum number of bins

observed from the histogram of with respect to odd

columns (white-runs) only.

Some of the assumptions concerning the state space search

(tunnel or traversal or ) are given below:

a. and are finite.

b. There must be at-least one between and

c. One agent (or robot) is employed at a time for tracing

.

d. There is no self-loop in (a cycle of length one).

e. Tunnel (Move to next state) if exist on .

Figure 5 shows the terminal points of and in an

uncompressed version for better understanding. It depicts 10

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 254

source nodes and 10 target nodes. The one-to-one function for the

figure 5 is defined below:

Based on the observation, the source does not have a

corresponding target . Therefore, an agent starting from may

reach or , where is presumed as a new (correspondence)

or . This infers that the agent may reach closer or nearer to

but not exactly the target , which is illustrated in the following

section.

Fig. 5. Line Separators displaying S and T at both ends of a document in

uncompressed documents. (Reference: A portion of ICDAR13 test image

- 320.tif)

4. Text-line Segmentation

Grid ( ) of a document image starts with a column of white-

run(s). The starting column of is referred [13] as a white space

that exists at the beginning of a text document. A larger depth

(white-runs) in this column indicates the text-line separation gap

(bands) along left margin of the document image between the

corresponding adjacent text-lines. Mid points are chosen as

terminal points (Source nodes) from the constructed bands

(corollary ) as shown in the figure 5. In this context, these points

do not necessarily possess larger white-runs. Therefore, the agent

needs to choose the starting point S among its vertical neighbors

having largest white-runs among its vertical neighbors within the

range from its current position. Now, assuming that the agent’s

start position S is fixed based on the corollary and this state is

at an initial search state . Next, the agent reaches at a

station or hub ( ) starting from resulting in a new search state

. An edge connecting between these two stations and

is the shortest path travelled by the agent to cover maximum

distance and obviously heading towards the probable target. The

distance or weight, from to an intermediate hub is

computed directly from the position, say , then it is

inferred in the first column of . The search process continues

until the agent reaches the corresponding or new target

(probable destination).

The basic idea is that for every , the agent needs to choose the

next hub which is nearer to or by selecting the white-runs

from the current position in and thus ensures that a direct edge

exists between the two stations (hubs). In other words, the

agent selects the largest white-runs from to adhoc, to navigate

to the next hub closer to or . At every , greedy strategy is

employed to identify the next successive hub h’ by choosing

largest white-runs (or largest weight) from the current and

henceforth the agent reaches or using a minimum number of

with respect to the hubs visited. The agent hopping hubs

starting from to or results in creating a path . is based

on the edges connected between the intermediate hubs starting

from to or and this indicates a progressive segmentation of

the two-adjacent text-lines in a document image, starting from the

left margin and precedes towards the right direction. If the total

search space for a given is one, then this indicates that the

agent has visited only two hubs and those are and

respectively; this articulates direct segmentation without much

variations exist because of skew, curvy and touching lines in the

handwritten document image. This situation sometimes would be

pronounced in-between two paragraphs in the document image

holding a large horizontal space between them and may occur in

the top and bottom of the documents which is entitled to have an

adequate margin spaces. However, the whole operation is carried

out with reference to .

Consider that the agent reaches an intermediate hub which is at

, then we perform local search by employing corollary to

reach to the next hub ( ). This situation is related to Hill-

Climbing [27] technique in Artificial Intelligence. If the hubs h

and h’ visited by the agent are located in the positions say

and respectively, then the search space (local

maxima) is written as.

The selection of each successive hub is based on the distance d

computed by summing up the entries from successive columns in

along axis starting from the first column with respect to the y

coordinate position. The distance computation is carried out for

every coordinate position which varies within the range of

from either side of position, , as given in the above equation.

The job of the agent is to select the largest weight (distance)

among these computed cumulative values which crosses the

current hub (intermediate hub). However, measuring the

distances for every SS ranging between to (both

inclusive) along axis is computationally expensive. In

computing, memoization [28] or memorize [28] is an

optimization technique used primarily to speed up computer

programs by storing the results of expensive function calls and

returning the cached result when the same inputs occur again. To

avoid this re-computation for every , we use memoization or

memoize technique in dynamic programming strategy to

remember the last computed of previous which in-turn

reduces the computational complexity. The gap or distance

between two successive hubs must be within a minimum

distance of a given threshold . is chosen heuristically based on

the experiments conducted on the benchmark datasets.

A common issue encountered as mentioned in the corollary is

when two adjacent text-lines are touching one another. In this

case, we just bypass or crossover the lines [29]. Algorithmically,

we choose next successive column (black-runs) with respect to

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 255

the current position. In other case, the same strategy is extended

when the ascenders and descenders appear to be greater than the

given threshold value , where the agent would not be able to go

beyond the search limit, say . One of the important issues

mentioned in corollary is when a distance between

hubs is greater than . is given below.

t’= maximum(histogram(values in odd colums of G))

where values in odd columns of G represents white runs

In this case, we calculate t’ by considering the largest bin from

the histogram of G with respect to the counts of white runs (odd

columns). If the distance is found to be larger than t’, then we

apply both the concepts namely ‘don’t care condition’ [29] and

‘minimal cut-sets’ [30] from graph theoretical domain. Therefore,

with a minimal number of crossovers, the agent bypasses and

reaches the next successive hubs. Finally, the agent will update its

knowledge base about each visited hub with reference to G

covering the distance from S to T or T’. Below we present the

proposed algorithm for text-line segmentation using .

Algorithm: Finding text-line segmentation in G using tunneling

approach

Input: Compressed representation G of a document image,

Terminal points – Source Nodes (Sn) at ever line terminals.

Vertical neighboring search threshold t.

Output: Text-line segmentation in G=TS

1. For-each Source Node (S) from Sn, with the position G(y’,x’), where, x’=1;

CurrentSum=0; Maximum=0; GridY=y’

While j=G(y’-t,x’) to G(y’+t,x’) the agent vertical search range is

between y’-t to y’+t (both inclusive), where j>=1 and j<=high of G.

a. Calculate: Cumulative Sum (Initialize Csum=0) for each j or

Use Dynamic Programming. Use Memoize to check data in

Knowledge base (KB)

a. Find Maximum and update coordinate position:

2. Update and coordinate position of :

The positions chosen in the grid is update

in

Goto Step 2 until

3. Stop: When the agent covers the distance .

4.1. Time Complexity for Text-line Segmentation

The following provides the time complexity estimated for the

proposed algorithm to identify the text-line segmentation

positions in compressed representation of a document image. The

worst-case scenario is calculated as where and

represent the number of rows and number of columns of the grid

respectively.

Usage of memoize technique [28] in dynamic programming

strategy has reduced to the complexity from

, that is which is a non-

deterministic polynomial.

For comparative study, the time complexity is also estimated for

a document image in its uncompressed format. For this, the

worst-case scenario is where and represent the

height and width of an uncompressed image respectively. It is

noted that is equal to whereas is lesser than .

Time complexity of spotting the line terminal points mentioned in

the literature [13] is estimated as in the worst-case

scenario. Therefore, the complexity for the proposed algorithm is

computed by cascading the two algorithms which result in

. Finally, the proposed method

takes .

The overall computation is tabulated (Table 2).

Table 2. Time Complexity

Complexity

Worst Case

Compressed

Domain

Uncompressed

Domain

Spotting Line Separators

Text-line Segmentation

without dynamic

programming

Text-line Segmentation

(Proposed method)

Overall Efficacy

The efficacy of the proposed algorithm for both compressed and

uncompressed versions is shown in Figure 6. The compressed

version requires less time, when compared to that of

uncompressed images for text line segmentation.

Fig. 6. Comparative analysis of both uncompressed and compressed

version of document images (Ref: ICDAR’13 test datasets)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 256

4.2. Illustration

In this section, we illustrate the working principle of Tunneling

algorithm using an example. The coordinate positions are tracked

by the agent (or robot) for text line segmentation with respect to

is shown in Table 3.

Table 3. RLE coordinate positions (Reference: A portion of ICDAR13

test image – 320.tif – text-lines 21 and 22)

Hubs

(Stations) Y-Coordinate X-Coordinate

Distance From

Source Node

Source 1977 1 767

Hub 1 1984 1 1058

Hub 2 2000 9 1295

Hub 3 2020 59 1325

Hub 4 2014 53 1626

Hub 5 2029 67 1901

Hub 6 2049 77 2085

For better understanding, an experimental result is shown in the

uncompressed version as well. Figure 7 shows segmentation of

two text-lines of a document image. Figure 8 illustrates the text-

line segmentation depicting transition hubs and a path tunneled

by an agent.

Fig. 7. Segmentation of two text lines depicted in uncompressed

documents (Reference: A portion of ICDAR13 test image – 320.tif – text-

lines 21 and 22)

Fig. 8. Segmentation of two text lines depicted in uncompressed

documents (Reference: A portion of ICDAR13 test image – 320.tif – text-

lines 21 and 22)

From the example, the agent starts from the Source Node ( ) and

reaches the Target Node ( ) visiting six intermediate hubs such

as Hub 1, Hub 2, Hub 3, Hub 4, Hub 5 and Hub 6. Here, the

source node is identified in a position, say . The agent

predicts that the Hub 1 in the position could be an

effective source node than the initial source [13] as articulated in

corollary . This is because the weight (distance) of Hub 1 is

larger when compared to the initial source as illustrated in table 3.

The total distance covered by the agent is 2085 units. The

threshold is chosen as 20 units and the total computation to

cover the distance along coordinates in both the direction by the

agent ranges from to

, i.e., to

respectively. We employ memoization technique to

reduce the complexity. Therefore, we make use of a knowledge

base ( ) which would be empty initially. Once we compute the

distances for both Source and Hub 1, we feed all the weights

(distances) ranging from to ,

i.e., to respectively into the KB. The size

of the KB would be a column size of . The triggers the agent

to choose next hub as Hub 2 as a successor of the Hub 1 which

covers the distance of 1295 units and perhaps the distance is

greater than the distance of current hub, that is the Hub 1 possess

1058 units. Next, agent archives the information from the KB to

avoid the total computation.

4.3. Choosing of Source Node

In this section, we introduce a mechanism to find the start node.

Figures 9 (a) and 9(b) show a sample image and its compressed

representation in respectively. Here, the source node is

not an optimal start point as mentioned in corollary . If we

consider the threshold for illustration purpose, then the

search scope will be of range covering from to

. Though the position is extended beyond the

grid boundary as described in the algorithm in step 2, we re-

define the search range starting from . Because of , the

two positions such as and are identified as the

largest among the search range. In this case, we take the source

node that is in the position than latter, because

is nearer or closer to the initial source node which is

. Finally, we choose as a start node, and thus the

distance covered by the agent would be 7 units and holding a

largest weight covering maximum distance and naturally

precedes or .

(a) (b)

Fig 9. Visualization of a search space predicting the actual source node.

Fig (a) represents the spatial coordinate of an image and Fig (b) represents

its RLE format. The green tile is the initial start node and the blue tile is

the actual start node. A red line represents the distance from source node

to the hub.

4.4. Handling Obstacles

(a)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 257

(b)

Fig. 10. Visualization of a search space tunneling an obstacle. Fig (a)

represents the spatial coordinate of an image and Fig (b) represents its

RLE format. The green tile is the initial start node and the blue tile is the

actual start node. Blue tiles represent the intermediate hub and red tiles

represents obstacle. A red line represents the distance from source node to

the hub.

This section provides the working principle of handling obstacles.

We encountered two types of obstacles while tunneling path as

mentioned in corollary e. Generally, an obstacle occurs due to

unconstrained writing style especially while scribing on an un-

ruled paper. One of the possibilities is when two adjacent lines

are touching one another at some positions. The other possibility

would be when the ascenders or desceders of the characters

extended beyond the search limit or range of the given threshold

( ). For illustration purpose, we have considered an example

which is shown in Figure 10(a) and Figure 10(b) representing

spatial domain and its compressed version of a document image.

We have chosen the threshold ( ) for demonstration. In this

example, an initial source node is assumed, and it holds a

position, . Previous section detailed about choosing

appropriate starting node and applying that concept results in a

new position, say as a new Start Node. The new source

node has a maximum weight or distance (say 7 units)

and falls within the range of , which is closer to the initial source

node . Further, the first successor hub is chosen as

which carries a weight of 6 units and covers the distance

of 11 units from the left margin. It is chosen based on fact that the

distance, say 11 units, is greater than the distance (weight) of

source node that covers the distance of 7 units. The other facts

include the search space (calculated distance) between the range

and is either lesser or equal to the

current position and this node aligns closer to its predecessor

node along axis. Further, next hopping hub is complicated

because we encounter an obstacle along the coordinate. Further,

the chosen threshold value ( ) is restricted within the search

space. In this situation, we can crossover the hurdle by choosing

the next hub as which is identified as a successor hub.

4.5. Finding Correspondence between Source and Target

Identification of correspondence between the terminal points

residing at opposite margins is one of the challenging aspects as

mentioned in the study [13]. In this article, actual source node is

chosen based on the weight (distance) which is distributed across

its neighbors. The destination is based on the search space of the

proposed model. The agent may start at a new source position and

would reach the probable destination, and this articulates the

correspondence between these two nodes present at extreme ends

of the document. This is applicable for all the source nodes and

target nodes. Figure 11 shows the correspondence between the

source and terminal nodes and the text-line segmentation. In

section 3 we have described the correspondence of the nodes

notably having one-to-one relation and strictly no onto

relationship and additionally no crossovers are allowed between

the terminal points. This could be related to the bi-partite graph

theoretical approach and thus maintaining the order or position.

Fig. 11. Line Terminals S and T at both ends of a document in

uncompressed documents (Reference: A portion of ICDAR13 test image -

309.tif)

The problem of over separation mentioned in the study [13]

highlights the aspect of spotting two or more separator points

within two adjacent text-lines. This occurs when a text line is

identified as a non-text (white space) region. One of the reasons

for over separation is because of concavity of the character. The

other affecting factor could be multiple disjoint fractions or

components which compose a character. Figure 12 depicts the

problem of over separation where source nodes and start

between two adjacent text-lines.

Fig. 12. Line Terminals S and T at both ends of a document in

uncompressed documents (Reference: A portion of ICDAR13 test image -

309.tif)

Taking an average distance between the separator points may not

produce the expected result as experimented in the research [13].

Our proposed method tackles this problem by understanding

correlation between the adjacent paths. In this example, we have

two source points such as s1 and s2 and the paths traced by the

agent reaches the terminal points and respectively. It is

evident that the terminal points and are very much closer to

one another. Even both the paths traced are aligned together

along coordinates most of the time. To resolve this issue, we

need to ignore the path(s) which relate to more hubs (nodes). In

other words, it is to retain the path which holds less intermediate

hubs. It is also necessary to retain a path which is relatively

parallel to both of its predecessor and successor paths.

Sometimes, this analogy may fail when a text line occupies less

space when compared to its predecessors as shown in figure 13.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 258

Fig. 13. Line Terminals S and T at both ends of a document in

uncompressed documents (Reference: A portion of ICDAR13 test image -

309.tif)

This occurs mostly with a last text-line of a paragraph having less

number of words. However, this situation is addressed by

understanding the average distance between the two adjacent

paths as mentioned above. It is computed by

We illustrate the distance between the adjacent paths with an

example. Figure 14 shows the paths traced along the two text-line

gaps.

Table 4 shows the coordinate positions of paths traced along

axis. We assume as 500 for illustration purpose. For every

interval , we take the coordinate positions for both paths.

Table 5 shows coordinate positions for the paths with an

interval gap of . The distance calculated for the paths (path 1

and path 2) is given under.

Fig. 14. Two paths traced along the text-lines

Table 4. RLE coordinate positions for two adjacent paths.

Path 1 Path 2

Hubs Y-Coordinates Distance

covered from

left margin

Y-Coordinates Distance

covered from

left margin

Source 1977 767 2079 186

Hub 1 1984 1058 2059 367

Hub 2 2000 1295 2073 682

Hub 3 2020 1325 2087 1132

Hub 4 2014 1626 2105 1389

Hub 5 2029 1901 2123 1665

Hub 6 2049 2085 2132 1870

Hub 7 - - 2124 2085

Table 5. RLE coordinate positions for two adjacent paths.

Hubs t’ Y-Coordinates

(Path 1), u

Y-Coordinates

(Path 2), v

Source 1-500 1977 2059

Hub 1 501-1000 1977 2073

Hub 2 1001-1500 2020 2105

Hub 3 1500-2000 2029 2132

Hub 4 2001-2500 2049 2124

5. Experimental Evaluation

5.1. Datasets

Our proposed method is evaluated on various benchmark

handwritten datasets such as ICDAR’13 [31] and others [17]

comprising of Kannada, Oriya, Persian and Bangla documents.

For experimental purpose, the compression standards for these

datasets are preserved as discussed in one of the studies [16].

5.2. Results

Our proposed method is evaluated based on two factors that we

came across in a study [31] - (i) Detection Rate ( ) and (ii)

Recognition Accuracy ( ). and are defined as follows:

The DR in spotting the separator points at both the left side and

the right side of the document is provided in literature [13] and

also been tabulated (Table 6).

Table 6. Detection Rate

Datasets

(Handwritten)

Total

Lines (N)

Detected

o2o Rate (%)

Left Right Left Right

ICDAR13 [31] 2649 2578 2502 97.31 94.45

Kannada [20] 4298 4173 4082 97.09 94.97

Oriya [20] 3108 3012 2911 96.91 93.66

Bangla [20] 4850 4650 4598 95.87 94.80

Persia [20] 1787 1690 1723 94.57 96.41

Table 7 shows the result of the proposed model applied on the

benchmark datasets. Our method focuses only on a terminal point

that is spotted on the left side of the document, so the RA for

segmenting the text line, entirely depends upon the separator

points spotted along left margin of the document. Figure 15

shows the comparative performance analyses of both DR and RA

for various benchmark datasets. It is evident that the algorithm

works better in the case of ICDAR datasets. We also witness that

the RA for the dataset of Persian script is lower when compared

to other datasets. This is because of the reason that the Persian

writing style starts from right end of the document and precedes

towards left direction unlike most of the other scripts. The other

reasons include more concavity and disjoints in the composition

of the character.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 259

Table 7. Accuracy Rate

Datasets Total Lines

(N)

o2o (Left) Recognition Rate in

% (Segmentation)

ICDAR13 2649 2578 89.2

Kannada 4298 4173 87.21

Oriya 3108 3012 85.00

Bangla 4850 4650 84.91

Persian 1787 1690 82.08

The algorithmic modeling deals with choosing a common

threshold value ( ) for every search space to handle the

obstacle is based on experimentation conducted on the datasets.

Figure 16 shows the RA for different threshold values with

respect to various datasets. The threshold value with 20 units

provides better accuracy rate with respect to ICDAR13, Kannada,

Oriya and Bangla. Whereas the threshold value possesses a unit

value of 28 which elevates the accuracy rate in case of Persian

script. In this research work, we have chosen the threshold value

as 20 units which is common to all the datasets.

Fig. 15. Performance evaluation with respect to different datasets

Fig. 16. Selection of threshold based for efficacy of the proposed model.

Real-time performance is measured employing the algorithm on

various benchmark datasets to understand the working principle

of the model. Figure 17 shows the comparative study of both (i)

compressed and (i) uncompressed versions of dataset. Here, we

witness that the processing time of compressed version of

document takes lesser time unit (milliseconds) when compared to

uncompressed version of documents. We also observe that there

is an invariable exponential increase in the amount of time for

performing text-line segmentation in case of uncompressed

documents with increase in the number of images.

(a)

(b)

(c)

(d)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 260

(e)

Fig. 17. Measurement of comparative real-time performance of both

compressed and uncompressed versions of the images. Time Unit is given

in milliseconds along y axis and x axis represents the sequence of images

of each dataset.

Finding correspondence between and is a challenging task.

However, we have attempted to find the correspondence based on

the correlation between the adjacent paths as illustrated in Section

4.5. The first of the document starts from the very first terminal

point identified in the first column of . is presumed as a

reference line. Now, the correlation is calculated between a new

path and the reference line with an interval of . The same is

repeated for the following paths with reference to source nodes.

Figure 18 shows the RA because of the model.

Fig. 18. RA after identifying the correspondence between the adjacent

text-lines with an interval of t’.

6. Conclusion

In this paper, we have proposed an algorithm that segments the

text-line of a document image by operating directly in its

compressed representation. We make use of current state-of-the-

art technology that identifies the terminal positions at every text-

lines in the compressed representation. We have shown the

working principle of the algorithm with an illustration. An agent

(or robot) has been involved to tunnel the path between the

terminal points spotted at the opposite extreme ends of the

document representation. We have discussed the search space for

tunneling the path when it encounters obstacles because of

oscillation, tilt, touching and curvy text-lines. We have advocated

the reasons for using the threshold values with respect to search

space by experimental results. Comparative study of processing

uncompressed and compressed versions for text-line

segmentation is also been detailed. Further, a significant

reduction in time complexity when compared to processing

uncompressed image for segmentation is showcased. Some of the

interesting observations are addressed to overcome the limitations

[13] such as finding the correspondence between the terminal

points. Interestingly, the tunnel or path traced that results in

segmenting two adjacent text lines are entirely based on the

terminal points. The source terminal points are predicted based on

the search space, SS, whereas the corresponding target points are

identified by the agent.

Though working directly on an uncompressed version of the

document is very challenging, we have designed a model that

operates directly on compressed representation of the document

images for text-line segmentation. We could achieve the

recognition accuracy of 89.2% using the benchmark dataset

(ICDAR13). The confidence level in achieving 89.2% with

respect to this dataset is 100%.

The proposed model has some of the limitations such as working

with invariable skew or tilt levels. The accuracy rate depends

upon the detection rate observed from the extensive literature and

a common threshold value t for entire datasets. Other limitation

includes calculating the correlation between the paths based on

reference line. One of the future avenues include working on bi-

directional search where we employ two agents starting

simultaneously from opposite terminal points and meets at a

junction. Employing multi-agents leads to parallel processing

which would improve the performance of the system. We have

followed unguided medium to tunnel the path by adapting various

algorithmic strategies including greedy and dynamic

programming with AI techniques. We could possibly avoid

wrong segmentation by using guided medium [4] along with the

strategy which would be one of the future direction of this

research work.

References

[1] Ahmed R, Al-Khatib WG, Mahmoud S (2017) A survey on handwritten documents word spotting. Int J Multimed Inf Retr

6(1):31–47

[2] Rath TM, Manmatha R (2007) Word spotting for historical documents. IJDAR 9(2–4):139–152

[3] Wshah S, Kumar G, Govindaraju V (2012a). Multilingual word

spotting in offline handwritten documents. In: ICPR, pp 310–313

[4] Wshah S, Kumar G, Govindaraju V(2012b) Script independent

word spotting in offline handwritten documents based on hidden

markov models. In: ICFHR, pp 14–19 [5] Yong X, Guangri Q, Yongdong X, Yushan S (2010) Keyword

spotting in degraded document using mixed ocr and word shape

coding. In: IEEE international conference on intelligent computing and intelligent systems, pp 411–414

[6] Kia O (1997) Document compression and analysis. PhD thesis, Institute for Advanced Computer Studies, University of Maryland

[7] Sayood K (2012) Introduction to data compression, 4th edn.

Morgan Kaufmann, Burlington [8] Salomon D, Motta G, Bryant D (2010) Handbook of data

compression. Springer, London

[9] T.4-Recommedation (1985) Standardization of group 3 facsimile apparatus for document transmission, terminal equipments and

protocols for telematic services, vol. vii, fascicle, vii. 3, Geneva.

Technical report [10] T.6-Recommendation (1985) Standardization of group 4 facsimile

apparatus for document transmission, terminal equipments and

protocols for telematic services, vol. vii, fascicle, vii. 3, Geneva.

Technical report

[11] Miano J (1999) Compressed image file formats: JPEG, PNG, GIF,

XBM, BMP. ACM Press, New York [12] Mohammed Javed, P. Nagabhushan & Bidyut B. Chaudhuri.

(2017). A review on document image analysis techniques directly

in the compressed domain. Artif Intell Rev. s10462-017-9551-9. Springer Science+Business Media Dordrecht 2017.

[13] Amarnath R & P. Nagabhushan. (2017). Spotting Separator points

at Line terminals in Compressed Document Images for Text-line Segmentation, International Journal of Computer Applications,

Volume 172 – Number 4

[14] Mukhopadhyay J (2011) Image and video processing in compressed domain. Chapman and Hall/CRC, Boca Raton

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 251–261 | 261

[15] Adjeroh D, Bell T, Mukherjee A (2013) Pattern matching in

compressed texts and images. Now Publishers. [16] Mohammed Javed, Krishnanand S.H, P. Nagabhushan, & B. B.

Chaudhuri. (2016). Visualizing CCITT Group 3 and Group 4 TIFF

Documents and Transforming to Run-Length Compressed Format Enabling Direct Processing in Compressed Domain International

Conference on Computational Modeling and Security (CMS 2016)

Procedia Computer Science 85 (2016) 213 – 221. Elsevier. [17] Alireza Alaei, P. Nagabhushan, Umapada Pal, Fumitaka Kimura,

"An Effcient Skew Estimation Technique for Scanned

Documents:An Application of Piece-wise Painting Algorithm", Journal of Pattern Recognition Research (JPRR), Vol 11, No 1

doi:10.13176/11.635 (2016).

[18] Amarnath R, P. Nagabhushan and Mohammed Javed, (2018) Line Detection in Run-Length Encoded Document Images using

Monotonically Increasing Graph Model, Accepted in 2018

International Conference on Advances in Computing, Communications and Informatics (ICACCI), September 19-22,

2018, Bangalore.

[19] Amarnath R, P. Nagabhushan and Mohammed Javed, (2018) Enabling Text-line Segmentation in Run-Length Encoded

Handwritten Document Image using Entropy Driven Incremental

Learning”, Third International Conference on Computer Vision & Image Processing (Accepted in CVIP-2018) September 29-October

01, 2018, IIIT Jabalpur, MP, India.

[20] N. Arvanitopoulos Darginis and S. Süsstrunk. Seam Carving for Text Line Extraction on Color and Grayscale Historical

Manuscripts. 14th International Conference on Frontiers in

Handwriting Recognition (ICFHR), Crete, Greece, 2014. [21] Zeyad Abd Algfoor, Mohd Shahrizal Sunar, & Afnizanfaizal

Abdullah. (2017). A new weighted pathfinding algorithm to reduce

the search time on grid maps. Expert Systems with Applications: An International Journal Archive Volume 71 Issue C, April 2017.

Pages 319-331

[22] Ronse C, Devijver P (1984) Connected components in binary images: the detection problem. Research Studies Press, Letchworth

[23] Breuel TM (2008) Binary morphology and related operations on

run-length representations. In: International conference on computer vision theory and applications - VISAPP, pp 159–166

[24] Hinds S, Fisher J, D’Amato D (1990) A document skew detection method using run-length encoding and the hough transform. In:

Proceedings of 10th international conference on pattern

recognition, vol 1, pp 464–468 [25] Shima Y, Kashioka S, Higashino J (1989) A high-speed

rotationmethod for binary images based on coordinate operation of

run data. Syst Comput Jpn 20(6):91–102 [26] ShimaY, Kashioka S, Higashino J (1990) A high-speed algorithm

for propagation-type labeling based on block sorting of runs in

binary images. In: Proceedings of 10th international conference on pattern recognition (ICPR), vol 1, pp 655–658

[27] Andrés Cano, Manuel Gómez-Olmedo, Serafín Moral, Joaquín

Abellán: Hill-climbing and branch-and-bound algorithms for exact and approximate inference in credal networks. Int. J. Approx.

Reasoning 44(3): 261-280 (2007)

[28] Peter Norvig. Paradigms of Artificial Intelligence Programming 1st Edition. Case Studies in Common Lisp.28th June 2014

[29] Nazih Ouwayed, Abdel Bela¨ıd. Multi-Oriented Text Line

Extraction from Handwritten Arabic Documents. 8th IAPR International Workshop on Document Analysis Systems - DAS’08,

Sep 2008, Nara, Japan. pp.339-346, 2008.

[30] Zhayida Simayijiang & Stefanie Grimm “Segmentation Graph-Cut”.

[31] Nikolaos Stamatopoulos, Basilis Gatos, Georgios Louloudis,

Umapada Pal, & Alireza Alaei Proceeding. (2013). ICDAR '13 Proceedings of the 2013 12th International Conference on

Document Analysis and Recognition Pages 1402-1406

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 262

Knowledge Discovery on Investment Fund Transaction Histories and

Socio-Demographic Characteristics for Customer Churn

Fatih CIL1, Tahsin CETINYOKUS*2, Hadi GOKCEN2

Accepted : 11/12/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: The need of turning huge amounts of data into useful information indicates the importance of data mining. Thanks to latest

improvement in information technologies, storing huge data in computer systems becomes easier. Thus, “knowledge discovery” concept

becomes more important. Data mining is the process of finding hidden and unknown patterns in huge amounts of data. It has a wide

application area such as marketing, banking and finance, medicine and manufacturing. One of the most commonly used application areas

of data mining is recognizing customer churn. Data mining is used to obtain behavior of churned customers by analyzing their previous

transactions. In the same manner using with obtained tendency, other active customers are held in the system. It is possible to make by

various marketing and customer retention activities. In this paper, it is aimed to recognize the churned customers of a bank who closed

their saving accounts and determine common socio-demographic characteristics of these customers.

Keywords: Data mining, Customer churn, Decision trees and classification rules, Mutual funds

1. Introduction

After the increase in the amount of data about customer’s

transactions and escalating competition in the industries, firms

have attempted to use their existing data efficiently in order to

obtain information which is used to understand purchasing and

service characteristics of customers. Due to increasing competition

and improvements in IT Technologies, obtained information will

provide advantage to firms, so it has been gaining importance.

The increasing usage of database systems and their extraordinary

increase in volume have made firms to be faced with the issue how

firms will derive benefit from these data. Traditional query or

reporting tools are inadequate in the face of massive amounts of

data. This is why the new tendencies and Data Mining is a result

of this search [1].

In this study, it is focused on two objectives. The first of these is

to learn the details of the transactions of the customers buying and

selling bank mutual funds including former customers who closed

their bank accounts after a stated transaction history. The second

objective is to obtain socio-demographic characteristics of the

customers who closed their investment account. With obtained

results, the rules are determined to prevent losing customers who

are tended to close their accounts. The main contribution is the first

study in literature considering mutual funds customer churn based

on data mining techniques.

It is organized as follows. In the second section, data mining and

classification algorithms used in the study are discussed. Then, a

general overview to bank mutual funds is presented in the third

section. In the fourth section, the study conducted is mentioned and

finally the results of the study are covered.

2. Data Mining, Decision Trees and Classification Rules

2.1. Data Mining

Data mining is a set of techniques that enable the extraction of

large volume data sets useful results. In problem solving, located

data in different sources or databases are analyzed. As a result of

this analysis, hidden images will emerge. Using the obtained

patterns can be given strategic management decisions.

Data mining is discovery driven. It involves various techniques

including statistics, decision trees, neural networks, genetic

algorithms and visualization techniques. Data mining has been

applied in many fields such as marketing, finance, banking, health

care, customer relationship management and organizational

learning [2].

Owing to the advancement of information technologies and global

competitiveness, data mining has increasingly become an

important and a widely known field. Data mining applications has

been varied recently, although exists in academic framework for

long years [3].

The major reason that data mining has attracted a great deal of

attention in the information industry in recent years is due to the

wide availability of huge amounts of data and the imminent need

for turning such data into useful information and knowledge. The

information and knowledge gained can be used for applications

ranging from business management, production control and market

analysis to engineering design and science exploration [1].

Data mining is searching of relations and rules for future prediction

by using software. Assuming that near future will not be very

different from today or past; rules obtained from existent data will

help us to form correct estimates of future trends.

Knowledge discovery in databases (KDD) is processing of data by

data mining techniques in databases. Processing large amounts of

data in data warehouses can be possible with new generation tools

and techniques. So, studies applied about KDD keep up to date.

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Operations Business Development at Finansbank İstanbul, TURKEY 2 Industrial Eng., Gazi University, Ankara – 06570, TURKEY

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 263

According to some sources; knowledge discovery in databases

refers to the overall process of discovering useful knowledge from

data, and data mining refers to a particular step in this process [4].

2.2. Decision Trees and Classification Rules

Classification data mining techniques are used when the dependent

(prediction) variable is discrete. Inferential techniques are used to

create decision trees and rules that effectively classify the dataset

as a function of its characteristics. Decision tree solution methods

provide automated techniques for discovering how to effect such

classifications [5,6]. At the top of the tree there is a root node, at

which the classification process begins. The test at the root node

tests all instances, which pass to the left if true, or to the right if

false. At each of the resulting nodes, further tests serve to continue

classifying the data. Each leaf of the tree represents a classification

rule. Rule-based solutions translate the tree by forming a conjunct

of every test that occurs on a path between the root node and the

leaf node of the tree [5]. One of the advantages offered by rule

induction algorithms is that the results may be directly inspected to

understand the variables that can be effectively used to classify the

data. Variables at the root node represent the strongest classifiers

(denoted as level 1 in the results section), followed by the next

strongest classifiers at each of the leaf nodes (denoted as level 2,

3, etc.).

In this study Id3 and J4.8 decision tree algorithm with PART, JRip

ve OneR classification rules are used. Id3 (basic divide-and-

conquer decision tree algorithm), J4.8 (implementation of C4.5

decision tree learner), PART (obtain rules from partial decision

trees built using J4.8), JRip (RIPPER algorithm for fast, effective

rule induction), OneR (1R classifier) (Witten and Frank, 2005).

3. Bank Investment Funds

The accumulation of all kinds of investors, whether large or small,

is managed by professional managers by distributing to the various

capital market instruments. These assets are called funds. Fund

portfolios consisting of treasury bonds, government securities,

repo, common stocks and other capital market instruments are

managed by professional portfolio managers in accordance with

internal regulations.

According to Capital Market Regulations investment fund

portfolios must be managed by institutions and portfolio

management companies. Investment funds are audited periodically

by Capital Markets Board and independent external audit firms.

Fund prices are determined by fund founder and these determined

prices are used next day in buying and selling funds.

Fund portfolio value is calculated in accordance with stock market

prices that funds are bought and sold. Total fund value is divided

by total share number that circulates on the day of valuation and

fund price is calculated.

Interests, dividends, trading gains and daily gains of portfolio

assets are recorded as income to fund at the same day and reflected

to the fund prices. Therefore, if an investor sells his/her funds,

he/she gets his/her share from the fund gaining

3.1. Discovery Studies on Investment Fund and Customer

Churn

Data mining are often used in studies about marketing, finance,

banking, manufacturing, health care, customer relations

management and organizational learning. Some of these studies

have been interested in the retaining the customers and minimizing

the customer churn.

Masand et al. [8] developed an automated system used historical

data of cellular phone customers. The system periodically

identifies churn factors for several statuses. In this study, decision

trees algorithms and neural networks algorithms were used.

Poel and Larivière [9] investigate predictors of churn incidence for

financial services as part of customer relationship management

(CRM) in their stuies. Their findings suggest that: demographic

characteristics, environmental changes and stimulating ‘interactive

and continuous’ relationships with customers are of major concern

when considering retention; customer behaviour predictors only

have a limited impact on attrition in terms of total products owned

as well as the interpurchase time.

Ahn et al., [9] investigated determinants of customer churn in the

Korean mobile telecommunications service market. In the study,

mediating effects of a customer's partial defection on the

relationship between the churn determinants and total defection

were analyzed and their implications were discussed. Results

indicate that customer's status change explains the relationship

between churn determinants and the probability of churn.

Ruta et al., [10] proposed a new k nearest sequence (kNS)

algorithm along with temporal sequence fusion technique to

predict the whole remaining customer data sequence path up to the

churn event in telecom industry.

Chu et al., [11] proposed a hybridized architecture to deal with

customer retention problems. In the system, common clustering-

followed- by- classification approaches have been used.

Aydoğan et al., [12] studied in a cosmetic firm to determine the

customer segment which tends to leave. So they developed

customized campaigns and marketing strategies for customer

profile who likely to leave. Clustering techniques were used for

segmentation and classification techniques were used for

determining the customer churn.

Telecommunications industry, customer churn is recognized as an

important research topic. In this view, Gopal and Meher [13] have

discussed the use of ordinal regression method to estimate the time

of customer churn and the duration of customer retention. On the

modeling of the customer churn, ordinal regression technic has

used in this study for the first time.

Farquad et al., [14] proposed a hybrid algorithm to predict churn

rates of customer using credit card. This algorithm is combination

of Support Vector Machine method and Naive Bayes method.

Huang et al., [15] presented a new set of features for broadband

internet customer churn prediction, based on Henley segments, the

broadband usage, dial types, the spend of dial-up, line-information,

bill and payment information, account information. The

experimental results showed that the new features with using these

four models (Logistic Regressions, Decision Trees, Multilayer

Perceptron Neural Networks and Support Vector Machines)

techniques are efficient for customer churn prediction in the

broadband service field.

Chen and Fan [16] proposed a framework of distributed customer

behavior prediction using multiplex data. Framework that is to

adapt to the emerging distributed computing environment, a novel

approach called collaborative multiple kernel support vector

machine (C-MK-SVM) is developed for modeling multiplex data

in a distributed manner, and the alternating direction method of

multipliers (ADMM) is used for the global optimization of C-MK-

SVM.

Zhang et al., [17] investigated the effects of interpersonal influence

on the accuracy of customer churn predictions and proposed a

novel prediction model that is based on interpersonal influence and

that combined the propagation process and customers’

personalized characters.

Slavescu and Panait [18] in their studies focuses on how to better

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 264

support marketing decision makers in identifying risky customers

in telecom industry by using Predictive Models. Based on

historical data regarding the customer base for a telecom company

they proposed a Predictive Model using Logistic Regression

technique and evaluate its efficiency as compared to the random

selection.

Devi and Madhavi [19] prepared a modeling study on purchasing

behavior of bank customers. In this study churn customers have

been predicted in India.

Jamal and Tang [20] in their study are directed for methods to

developing customer retention and loyalty strategies. One of them

is about calculating likelihoods of next action taken by customers.

Another one is calculating risk scores of customer churn based on

customer attributes.

A social network analysis for customer churn prediction was

improved by Verbekea et al. [21]. A new model setup was

introduced in order to combine a relational and nonrelational

classification model. It was shown that the model setup boosted the

profits generated by a retention campaign.

Farquad et al., [22], an analytical CRM (customer relationship

management) application was achieved with churn prediction

using comprehensible support vector machine. It was indicated

that the generated rules acted as an early warning expert system to

the bank management.

A churn prediction in telecommunication industry was improved

using data mining techniques by Keramati et al. [23]. Above 95%

accuracy for Recall and Precision was achieved with the proposed

methodology in addition, a new methodology was introduced for

extracting influential features in dataset. As can be seen in the

literature, there are no studies on the banking sector mutual fund

clients.

4. Application

Due to increasing competition and new emerging actors in banking

sector, banks’ efforts for keeping customers and getting maximum

benefit from them increased.

Finding new customers, advertising and promotion costs to find

new customers increase day by day. Even though lost customers

are compensated by new customers, cost of keeping customers is

lower than cost of finding new customers.

The aim of this paper is to determine which transaction history and

socio-demographic characteristics of customers cause them to

close their accounts. With these findings determining the

customers that tend to close their accounts is aimed, as well. The

data of nearly 87.000 fund customers’ annual investment behaviors

and 65.525 customers’ socio-demographic characteristics are

considered. In the analysis, nearly 4.000.000 transaction of

investment funds data that belongs to these customers is used.

4.1. Preprocessing of Socio-Demographic Characteristics Data

of Customers

Socio-demographic characteristics data of 65.525 customers has

been taken from a bank’s database and preprocessed. The socio-

demographic characteristics data that has been used and

preprocessing procedure is as follows;

The living city information for customers has been taken from

bank’s data base as city codes. To make city codes categorical and

add development value of cities to the study, classification has

been made by using per capita GDP data of cities. According to

Bank Investment Funds Specialists, city variable must get

maximum 5 categorical values. Taking into consideration the Bank

Investment Funds Specialists’ views, k-means algorithm,

commonly used in classification, has been applied from 2 to 6 for

class number. Square errors are shown in table 1.

Table 1. Sum of square errors according to K-means algorithm (%)

Class

Number 2 3 4 5 6

Sum of

Square

Errors

0.25691 0.25688 0.25687 0.25686 0.25686

Considering the Table 1, it is seen that sum of square errors is

decreasing. Because Bank Investment Funds Specialists think that

city variable must get maximum 5 categorical values, class number

is determined as 5.

According to determined class number, 5, and value range for

classes, cities that customers live and categorical values for that

cities are shown in table 2.

Table 2. Cities and Categorical Values

Cities Per capita GDP Categorical

Values

Kırklareli, Yalova, Muğla,

İzmir, İstanbul, Zonguldak,

Ankara, Kırıkkale, Bilecik, Eskişehir, Bursa, Tekirdağ,

Manisa, İçel, Edirne

2339 - 4214 TL 1

Çanakkale, Antalya, Artvin, Denizli, Nevşehir, Sakarya,

Aydın, Karaman, Balıkesir,

Burdur, Rize, Kilis

1817 - 2338 TL 2

Kayseri, Kütahya,

Kastamonu, Niğde, Hatay,

Elazığ, Samsun, Çorum, Gaziantep, Karabük,

K.Maraş, Tunceli, Konya, Isparta, Trabzon, Kırşehır,

Sinop, Giresun, Amasya,

Uşak, Malatya, Sivas, Diyarbakır, Afyon, Batman,

Erzincan, Osmaniye, Düzce,

Çankırı, Siirt, Gümüşhane, Ordu, Erzurum, Bartın,

Bayburt, Şanlıurfa, Mardin,

Aksaray, Adıyaman

920 - 1816 TL 3

Kars, Van, Iğdır, Yozgat,

Ardahan, Hakkâri, Bingöl,

Bitlis, Şırnak, Muş, Ağrı, Tokat

∞ - 919 TL 4

Kocaeli, Bolu 4215 - ∞ TL 5

City variable has been used as a categorical variable in the study.

Table 3 gives the description of variables after preprocessing of

socio-demographic characteristics data of customers.

After 65,525 customer social- demographic data has been passed

through pre-processing stage as defined above, a matrix which has

65525 x 15 elements has been acquired.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 265

4.2. Pre-processing of Investment Fund Buy / Sell Transaction Data

For selected customers, about 4,000,000 investment funds buy /

sell transaction data have been filtered and listed for every single

customer in terms of date. Listed transactions have been summed

up in terms of months and transaction amount of every single

customer’s in these months have been determined and then

percentage changes on principal amount for July 20** of

transaction amounts have been calculated. For a sample customer,

investment fund buy / sell order data is given in table 4.

Table 4. For a sample customer, investment fund buy / sell order data

Cu

stom

er

Nu

mber

Cap

ital

July

20

**

Au

g. 2

0**

Sep

. 20*

*

Oct

. 20

**

No

v. 2

0**

Dec

. 20

**

Jan

. 20*

*

1 18759.72 -0.26 0 1.47 1.63 1.61 1.76 2.26

From variables in customer’s social – demographic properties, by

assessing “Liability Date” as “t-1” month, investment fund buy /

sell transaction order data of every single customer have been

reorganized.

For example, the customer having 1490 customer number made his

Table 3. Description of variables

Variable Description Range

Customer

numbers

Customer numbers are begun from 1 and given to all customers by increasing

one. Customer number variable has been used as numerical variable. 1- 65.525

Gender Gender information of customers has been taken from bank’s data base and

changed as “1” for “Male” and “2” for “Female”. Gender variable has been used

as a categorical variable in the study.

“1”, “2”

Age By using the date of birth information the age of customers has been determined

and identified age groups have been categorized.

younger than 20, “1” between 21 – 30, “2”

between 31 – 40, “3”

between 41 – 50, “4” between 41 – 50, “5”

older than 60, “6”

Marital status Marital status data has been changed as “1” for “Married” and “2” for “Single”. “1”, “2”

Educational

background

Educational background information for customers has been taken from bank’s

data base as “Primary School”, “Elementary School”, “Primary and Elementary

School”, “High School”, “Vocational School”, “Undergraduate”, “Graduate” and “Postgraduate” and changed in preprocessing.

Primary School, “1”

“Elementary School”, “2” “Primary and Elementary School”, “3”

“High School”, “4”

“Vocational School”, “5” “Undergraduate”, “6”

“Graduate”, “7”

“Postgraduate” “8”

City The living city information. “1”, “2”, “3”, “4”, “5”

House that customers

live

House information for customers has been taken from bank’s data base and has

been changed.

“Own house”, “1”

“Rental”, “2”

“Dig”, “3” “Belongs to family members”, “4”

Experience in the last job

Job experience (year) information for customers.

less than 1 year, “1”

between 2 – 5 years, “2” between 6 – 10 years, “3”

more than 11 years, “4”

Monthly net income

Net income of customer has been filtered.

< 1.000 TL, “1” 1.001 - 1.500 TL, “2”

1.501 - 2.500 TL, “3”

2.501 - 4.000 TL, “4” 4.001 TL >, “5”

Occupation Occupation type of customer information has been filtered.

Public Sector, “1”

Private Sector, “2” Self Employed, “3”

Supplementary Income Holder, “4”

Retired, “5” Student, “6”

Housewife, “7”

Unemployed, “8”

Spouse employee

Customer’s spouse employed status data has been changed as “1” for “Yes” and “2” for “No”.

“1”, “2”

Automobile of customer

Information that whether or not customer has got an automobile. “1” for “Yes” and “2” for “No”.

“1”, “2”

Capital TL value of customer’s total amount of investment fund.

Asset /

Liability

Information that whether or not the customer closed his / her investment

account

Closed accounts, “P”

Unclosed accounts, “A”

Liability date

The month in which customer made his / her account inactive. Liability date

variable has been used in order to organize investment fund buy / sell transaction table.

“07.20**”, “08.20**”, “09.20**”,

“10.20**”, “11.20**”, “12.20**” and “01.20**”

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 266

/ her account inactive in December 20**. So for this customer, “t-

1” month is December 20**, “t-2” month is November 20**, “t-3”

month is October 20**, “t-4” month is September 20**, “t-5”

month is August 20** and “t-6” month is July 20**. “t-7” cell was

filled as zero (0). Investment fund buy / sell order data of the

customer having 1490 customer number is given in table 5.

Table 5. Investment fund buy / sell order data of the customer having

1490 customer number

Cu

stom

er

Nu

mber

Capital t-7 t-6 t-5 t-4 t-3 t-2 t-1

1490 2642.78 0 -0.23 -0.02 0.31 -0.01 0 -1

In order to categorize percentages of organized investment fund

buy / sell order data, series of processed has been done. In order to

categorize zero values, not increased or decreased values, “S” code

meaning “fixed” has been given. In order to categorize decreased

percentage rates, clumping has been done.

According to opinion of bank investment fund specialists,

decreasing percentage rate values could be clumped into between

7 and 10 clusters and shouldn’t be more than 10 clusters. In this

process, regarding bank investment fund specialists’ opinion, k-

means algorithms used commonly in clumping have been used. K-

means algorithm has been implemented for number of clusters

between 7-10 and table 6. error squares have been got.

Table 6. Sum of k-means algorithm decreasing percentage values’ error

squares

Number of Clusters 7 8 9 10

Sum of Square Errors 78.79 73.46 59.72 55.77

When sum of error squares is analyzed, sharp decrease between

state of 8 clusters and 9 clusters has been seen (look Figure 1.) For

this reason, cluster number has been determined as 8.

According to determined number of clusters and cluster range

values, decreasing percentage rates have been categorized.

Categorizing process is built by gathering “D” meaning decrease

in percentage rates and clusters number. Information showing

which decreasing percentages belongs to which category is shown

in table 7.

Fig 1. Change in sum of k-means error squares graph (Table 6.)

Clumping has been also made in order to categorize increasing

percentage rates. Aim of this is to determine how many clusters

increasing values should be classified.

According to bank investment fund specialists’ opinion, increasing

percentages should be separated at most 4 clusters. Because they

think that as the study has been made on the customers who closed

their investment fund accounts, analyzing increasing percentages

is not useful for analysis. Table 8. has been acquired by

implementing k-means algorithm for between 2-4 clusters.

Table 7. Categorical classification of decreasing percentages

Categorical

Values

Range of Decreasing

Percentages Number of Values

D8 (-1.00) - (-0.81) 3409 (%5)

D7 (-0.80) - (-0.54) 3318 (%5)

D6 (-0.53) - (-0.35) 4781 (%7)

D5 (-0.34) - (-0.21) 6717 (%10)

D4 (-0.20) - (-0.11) 10045 (%15)

D3 (-0.10) - (-0.05) 14133 (%21)

D2 (-0.04) - (-0.02) 15794 (%24)

D1 (-0.01) - (0) 7818 (%12)

When the table analyzed, it is clear that sum of error squares is

continuously decreasing. According to bank investment fund

specialists’ opinion, increasing percentages should be separated at

most 4 clusters and for this reason number of clusters has been

determined as 4.

According to determined number and range of clusters, increasing

percentages have been categorized. Process of categorizing has

been built by gathering “Y” meaning increasing percentages and

clusters number. Information showing which increasing

percentages belongs to which category is shown in table 9.

Table 8. Error square sum of k-means algorithm decreasing percentage values

Number of

Clusters

2 3 4 5

Sum of Square

Errors

0.90 0.35 0.22 0.20

Table 9. Categorical classification of increasing percentages

Categorical

Values

Range of Increasing

Percentages Number of Values

Y4 ∞ - (2.00) 13086 (%4)

Y3 (2.00) - (0.77) 37448 (%11)

Y2 (0.76) - (0.26) 80658 (%25)

Y1 (0.25) - (0) 196452 (%60)

Investment fund buy / sell order data for categorized percentages

and customer having 1490 customer number is shown in table 10.

By gathering social demographic data from pre-processing stage

and investment fund buy / sell order data, a matrix which has 65525

x 9 elements has been acquired.

Table 10. Investment fund buy / sell order data for categorized

percentages and customer having 1490 customer number

Cu

stom

er

Nu

mber

Cap

ital

t-7

t-6

t-5

t-4

t-3

t-2

t-1

1490 2642.78 S D5 D2 Y2 D1 S D8

4.3. Investment Fund Buy / Sell Transaction Order Data Analysis

By this classification, it is aimed to determine transactions during

separation process by analyzing transaction details of customers

who did not want to make investment fund transaction anymore

and closed their accounts.

Customer investment fund buy / sell order data has been analyzed

by J4.8, PART, JRip, Naive Bayes and OneR classification

algorithms. In classification techniques, as object variable, “Asset

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 267

/ Liability” variable showing that customer has made his / her

account inactive has been used.

Decision trees, built by analyzing investment fund buy / sell order

data by classification algorithms have been analyzed. As PART,

JRip and OneR algorithms are rule-learner algorithms, their results

are also in rule forms. But since J4.8 gives results in the form of

decision tree, the decision tree acquired by J4.8 algorithm has been

converted into rule. After this conversion, 31 rules have been

acquired.

Analysis results of classification algorithms Number of Rules /

Number of Leaves, True Positive Rate (TP Rate), False Positive

Rate (FP), Precision are analyzed by Kappa statistics and

Confusion matrixes. Comparison table of analysis results have

been prepared and are given in table 11.

Table 11. Comparison of classification results of investment fund buy /

sell order data

Alg

ori

thm

Nu

mber

of

rule

s /

Nu

mber

of

elem

ents

of

dec

isio

n t

ree

Nu

mber

of

rule

s fo

r

pas

sive

cust

om

ers

Co

rrec

tly c

lass

ifie

d

per

centa

ges

fo

r p

assi

ve

cust

om

ers

No

t co

rrec

tly c

lass

ifie

d

per

centa

ges

fo

r p

assi

ve

cust

om

ers

Kap

pa

Sta

tist

ics

Nu

mber

of

no

t co

rrec

tly

clas

sifi

ed c

ust

om

ers

J4.8 123 31 95.8 0.1 0.9517 62

PART 79 25 95.6 0.1 0.9516 65

JRip 14 14 99.2 0.2 0.9515 12

OneR 1 1 38.5 0.2 0.5144 914

When classification results are analyzed, it has been seen that, the

highest number of rules are derived by J4.8 algorithm. Most

powerful ones of these rules are;

• Rule 1: If (t-2 = D8) and (t-1 = S) = P (492, % 33.13)

• Rule 2: If (t-7 = S) and (t-6 = S) and (t-1 = D8) = P (354, %

23.84)

• Rule 3: If (t-3 = D8) and (t-2 = S) and (t-1 = S) = P (212, %

14.28)

• Rule 4: If (capital > 85.95) and (t-4 = D8) and (t-3 = S) and (t-

2 = S) and (t-1 = S) = P (88, % 5.93)

• Rule 5: If (t-7 = S) and (t-6 = Y2) and (t-1 = D8) = P (71, %

4.78)

The rules built reveals transaction characteristics of the customers

who closed their accounts. These rules will be used to determine

potential customers who may be close their accounts in the future.

However, these rules do not reveal which social demographic

characteristics in which rules. Customer social demographic

characteristics can be included by a new analysis. In order to do

this, the customers obeying the rules have been determined and

customer social demographic data has been reorganized. Further

analysis has been made by reorganized social demographic data

and classification techniques.

4.4. Customer Socio-Demographic Data Analysis

Investment account bank mutual fund customers that have ceased

to be turning off customers analyzed with Id3, J4.8, PART and

JRip classification algorithms. The classification techniques as the

target variable that specifies the path that the customer's

investment account while making passive rule number is used as a

variable.

The Classification results of the analyzed algorithms were

analyzed criteria such as True Positive Rate (TP Rate), False

Positive, Kappa statistics and Confusion matrix. Analysis results

comparison table was created. The comparison chart is given in

table 12.

Table 12. The classification results comparison chart of customer socio-

demographic data

Algorithm

Percentage

of correctly classified

records

Percentage

of incorrectly

classified records

Kappa Statistics

Number of

incorrectly classified

records

Id3 98.04 1.95 0.9757 29

J4.8 50.84 49.15 0.3244 730

PART 57.64 42.35 0.4373 629

JRip 33.53 66.46 0.008 987

When the classification results were examined, the accuracy rate

of 98.04% was observed with the Id3 algorithm is the highest

percentage of correctly classified records of the algorithm.

The investment fund buying selling instruction data of investment

account bank mutual fund customers, who after a certain

transaction history has ceased to be turning off customers by

disabling their investment account and socio-demographic data

classification of results of two-stage decision tree has been

reached. In a bank investment fund performs but a certain

transaction history after the investment account to disable the bank

mutual fund customers to be out of the customers’ investment fund

buying selling instruction data and socio-demographic data

classification has resulted two-stage decision tree. The mutual fund

sales practice data involves the rules that shows after which motion

the customers closed down their investment account. The socio-

demographic characteristics of the customers were derived from

the classification of the clients’ socio-demographic characteristics.

4.5. The Combination of the Customer's Investment Account Closed Down with the Socio-Demographic Characteristics

Classification results for investment account closing transactions

of the customer done by J4.8 algorithm and classification results

for social – demographic characteristics of the customer have been

combined. With this combination, the decision tree having two

levels can be used as a decision tree having only one level. Some

of the decision tree leaves with highest explanation power is given

below as examples and the decision tree structure of Rule 1 is given

figure 2.

Rule 1: If (t-2 = D8) and (t-1 = S) = P (492, 33.13%)

Socio-demographic characteristics;

• (Occupation = 2)

• (Occupation = 3) and (Age = 2) and (Experience (year) in the

last job = 2) and (Gender = 1)

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 3)

and (Educational Background= 6)

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 3)

and (Educational Background= 8)

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 5)

and (Marital Status= 1)

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 5)

and (Marital Status= 2) and (Educational Background= 5)

• (Occupation = 3) and (Age = 4) and (Experience (year) in the

last job = 1) and (Method of Study= 3) and (Gender = 1)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 268

Fig 2. The decision tree structure of Rule 1

Rule 2: If (t-7 = S) and (t-6 = S) and (t-1 = D8) = P (354, % 23.84)

Socio-demographic characteristics;

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 2)

and (Gender = 1)

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 5)

and (Marital Status= 2) and (Educational Background= 7)

• (Occupation = 3) and (Age = 5) and (Experience (year) in the

last job = 2) and (Monthly Net Income = 2)

• (Occupation = 4) and (Gender = 1)

• (Occupation = 5) and (Age = 2)

The decision tree structure of Rule 2 is given figure 3.

Fig 3. The decision tree structure of Rule 2

Rule 3: If (t-3 = D8) and (t-2 = S) and (t-1 = S) = P (212, 14.28%)

Socio-demographic characteristics;

• (Occupation = 3) and (Age = 2) and (Experience (year) in the

last job = 1) and (City = 1)

• (Occupation = 3) and (Age = 4) and (Experience (year) in the

last job = 1) and (Method of Study= 2) and (Gender = 2) and

(Monthly Net Income = 3)

• (Occupation = 3) and (Age = 4) and (Experience (year) in the

last job = 3)

• (Occupation = 3) and (Age = 4) and (Experience (year) in the

last job = 4) and (Gender = 1)

The decision tree structure of Rule 3 is given figure 4.

Fig 4. The decision tree structure of Rule 3

Rule 4: If (capital > 85.95) and (t-4 = D8) and (t-3 = S) and (t-2 =

S) and (t-1 = S) = P (88, 5.93%)

Socio-demographic characteristics;

• (Occupation = 3) and (Age = 4) and (Experience (year) in the

last job = 2) and (House that customers live= 3)

• (Occupation = 4) and (Gender = 2)

• (Occupation = 5) and (Age = 5) and (Educational Background=

7)

The decision tree structure of Rule 4 is given figure 5.

Rule 5: If (t-7 = S) and (t-6 = Y2) and (t-1 = D8) = P (71, 4.78%)

Socio-demographic characteristics;

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 2)

and (Gender = 2)

• (Occupation = 3) and (Age = 3) and (Monthly Net Income = 3)

and (Educational Background= 7)

• (Occupation = 5) and (Age = 4) and (Experience (year) in the

last job = 2) and (Educational Background= 6) and (House that

customers live= 1)

Rule 1

Soldier Lawyer

Occupation

Age

21-30 31-40

Experience (year) in the last job

Monthly Net Income

2-5

Gender

Male

Rule 1

1.501-2.500 TL 4.001 TL >

Educational Background

Marital Status

Undergraduate Postgraduate

Rule 1 Rule 1

Married Single

Rule 1Educational Background

Senior High School

Rule 1

Experience (year) in the last job

41-50

Method of Study

< 1

Self Employed

Gender

Male

Rule 1

Rule 2

Bank Employee Lawyer

Occupation

Age

21-30 31-40

Monthly Net Income

1.001-1.500 TL 4.001 TL >

Marital Status

Male

Rule 2

Single

Educational Background

Graduate

Rule 2

Experience (year) in the last job

51-60

Monthly Net Income

2-5

1.001-1.500 TL

Rule 2

Bureaucrat

GenderAge

Male

Rule 2

Gender

Lawyer

Occupation

Age

21-30

Experience (year) in the last job

Monthly Net Income

<1

City

Group 1

Rule 3

<1 6-10

Method of Study

Private Sector

Rule 3

Experience (year) in the last job

41-50

Gender

11>

Male

Rule 3

Rule 3

Gender

Female

1.501 – 2.500 TL

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 269

• (Occupation = 8) and (Educational Background= 2) and (Age =

2) and (Monthly Net Income = 4)

The decision tree structure of Rule 5 is given figure 6.

Fig 5. The decision tree structure of Rule 4

Fig 6. The decision tree structure of Rule 5

5. Conclusion

The most common investment account closure rule is rule 1. By

following this rule, the percentage of customers that close their

investment account is 33.13%. According to the rule, the customer

who make transactions in line with D8 amount in t-2 month and

make no transactions in t-1 month in other words the customer who

sells at least 80% percentage of investment funds in t-2 month and

do not make any transaction in t-1 month, will be closed his/her

bank account in t month and will be evaded to be customer. If these

people has no transaction in t-1 month and closed their investment

accounts in month t, actually sold the entire investment funds in

their accounts in month t-2. The customers behave fast to sell funds

in their investment accounts and they only wait one month for

closing their investment accounts. The bank does not have much

time to contacting with these customers to convince them not to

close the accounts. Thus, special attention is needed for this group.

When we analysis socio-demographic data of these customers who

close their investment accounts by applying this rule, we reach

many different professions, different working types, and people of

different age groups. One of the interesting features that a customer

whose profession is 2 and 39 in other words soldiers and drivers is

acting according to this rule.

The rule 2 is one of the most common investment account closure

rule. By following this rule, the rate of customers who close their

accounts is 23.84%. According to the rule the customer that not

trading in t-7 and t-6 months, then trading in t-1 with D8 amount

and that client closes it investment fund account. The customers,

who close their investment accounts according to this rule, behave

quickly than the customers that close their accounts in line with the

rule 18. Similar to the rule 18, the bank does not have much time

to contact with the clients to convince them to not to close their

investment accounts. Special attention is required for this group.

When we analysis socio-demographic data of these customers who

close their investment accounts by applying this rule, we reach

many different professions, different working types, and people of

different age groups. The male customers whose profession is 4 in

other words Minister, Deputy Bureaucrat acting according to this

rule.

One of the most common investment rules to close an investment

account is the rule 3 with 14.28% percentages. According to the

rule, the customers operate in t-3 in line with D8 amount, not

operating in t-2 and t-1 and in month, t closes the account and

leaves being an investment fund client. This group waits one more

month to close their accounts with respect to the rules 18 and 17,

and this group also involves the customers who need special

interest.

The customers who close their accounts according to the rule 4 act

slowly than the customers that close the accounts according to the

rules 1, 2, and 3. Moreover, the customers making transaction in

month t-4 in line with D8 amount, and then wait in t-3, t-2, t-1

months with not making transactions and not closing their

accounts. The customers whose professions are 17 in other words

workers and the female customers whose profession is 4 in other

words Minister, Deputy Bureaucrat acting according to this rule.

This work supports to identify the customers’ socio-demographic

characteristics, specialties of investments and the customers who

turn to close the bank account can be determined. By this way,

promotional activities would be lead to behave proactively in order

not to lose the customers.

References

[1] Han, J., Kamber, M., Data mining: concepts and techniques 1st ed., Morgan Kaufmann, USA, 3-16, (2001).

[2] Chien C.-F., Chen, L.-F., Data mining to improve personnel

selection and enhance human capital: A case study in high-technology industry, Expert Systems with Applications, 34(1): 280-

290 (2008).

[3] Giudici, P., Applied data mining: statistical methods for business and industry 1st ed., John Wiley & Sons, England, 1-15, 85-110,

(2003).

[4] Fayyad, U., Piatetsky-Shapiro G., Smyth P., From data mining to knowledge discovery in databases, American Association for

Artificial Intelligence, 3(17): 37-54 (1996).

[5] Apte, C., Weiss, S., Data mining with decision trees and decision rules, Future Generation Computer Systems, 13, 197–210 (1997).

[6] Fernandeza, I.B., Zanakisa, S. H., Walczakb, S., Knowledge discovery techniques for predicting country investment risk.

Computers & Industrial Engineering, 43, 787–800. (2002).

Rule 4

Bank Employee Lawyer

Occupation

Age

51-60

Graduate

Experience (year) in the last job

41-50

House that customers live

2-5

Dig

Rule 4

Bureaucrat

GenderAge

Female

Rule 4Educational background

Bank Employee Lawyer

Occupation

Age

41-50 31-40

Monthly Net Income

1.001-1.500 TL 1.501-2.500 TL

Female

Rule 5

Educational Background

Graduate

Rule 5

Financial Advisor

Age

Gender

Experience (year) in the last job

2-5

Educational Background

Undergraduate

House that customers live

Own House

Rule 5

Educational Background

Elementary School

Age

21-30

Monthly Net Income

2.501-4.000 TL

Rule 5

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 262–270 | 270

[7] Witten, I., H., Frank, E., Data mining: practical machine learning

tools and techniques 2nd ed., Morgan Kaufmann, USA, 62-415 (2005).

[8] Masand, B., Datta, P., Mani, D.R., Li, B., CHAMP: A prototype for

automated cellular churn prediction, Data Mining and Knowledge Discovery 3, Netherlands, 219 - 225 (1999).

[9] Poel, D., and Lariviere, B., Customer attrition analysis for financial

services using proportional hazard models, European Journal of Operational Research, 2004, vol. 157, No 1, 196-217.

[10] Ahn, J.H., Han, S.P., Lee, Y.S., Customer churn analysis: churn

determinants and mediation effects of partial defection in the Korean mobile telecommunications service industry, Science Direct,

Telecommunications Policy, 30: 552 - 568 (2006).

[11] Ruta, D., Nauck, D., Azvine, B., K nearest sequence method and its application to churn prediction, Springer-Verlag, Berlin Heidelberg,

207 - 215 (2006).

[12] Chu, B.H., Tsai, M.S., Ho, C.S., Toward a hybrid data mining model for customer retention, Knowledge-Based Systems, 20 (8) (2007),

pp. 703–718.

[13] Aydoğan, E., Gencer, C., Akbulut, S., Churn analysis and customer segmentation of a cosmetics brand using data mining techniques,

Journal of Engineering and Natural Sciences, Sigma, 26 (2008).

[14] Gopal, R. K., Meher, S. K., Customer churn time prediction in mobile telecommunication industry using ordinal regression,

Springer-Verlag, Berlin Heidelberg, 884 - 889 (2008).

[15] Farquad, M.A.H., Ravi, V., Raju, S.B., Data mining using rules extracted from svm: an application to churn prediction in bank credit

cards, Springer-Verlag, Berlin Heidelberg, 390 - 397 (2009).

[16] Huang, B. Q., Kechadi, M. T., Buckley, B., Customer churn prediction for broadband internet services, Springer-Verlag, Berlin

Heidelberg, 229 - 243 (2009).

[17] Chen, Z.-Y., Fan, Z.-P., Distributed customer behavior prediction using multiplex data: a collaborative MK-SVM approach,

Knowledge-Based Systems, 35 (2012), pp. 111–119.

[18] Zhang, X., Zhu, J., Xu, S., Wan, Y., Predicting customer churn through interpersonal influence, Knowledge-Based Systems, 28

(2012), pp. 97–104.

[19] Slavescu, E., Panait, I., Improving customer churn models as one of customer relationship management business solutions for the

telecommunication industry, “Ovidius” University Annals, Economic Science Series, 2012, vol. 12, Issue 1/201.

[20] Devi P., U., Madhavi, S., Prediction of churn behavior of bank

customers using data mining tools, Business Intelligence Journal, 2012, vol.5, 96-101.

[21] Jamal, Z., Tang, H.K., Methods and systems for identifying

customer status for developing customer retention and loyalty strategies, United States, Patent Application Publication, (2013).

[22] Verbeke, W., Martens, D., Baesens, B., Social network analysis for

customer churn prediction, Applied Soft Computing, 2014, vol. 14, 431–446.

[23] Farquad, M.A.H., Ravi, V., Raju, S.B., Churn prediction using

comprehensible support vector machine:An analytical CRM application, Applied Soft Computing, 2014, vol. 19, 31–40.

[24] Keramati, A., Jafari-Marandi, R., Aliannejadi, M., Ahmadian, I.,

Mozaffari, M., Abbasi, U., Improved churn prediction in telecommunication industry using datamining techniques, Applied

Soft Computing, 2014, vol.24, 994–1012.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 271–274 |271

Improved Nelder-Mead Optimization Method in Learning Phase of

Artificial Neural Networks

Hasan Erdinc KOCER1, Mustafa Merdan*2, Mohammed Hussein IBRAHIM3

Accepted : 14/12/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: Artificial neural networks method is the most important/preferred classification algorithm in machine learning area. The weights

on the nets in artificial neural directly affect the classification accuracy of artificial neural networks. Therefore, finding optimum values of

these weights is a difficult optimization problem. In this study, the Nelder-Mead optimization method has been improved and used for

training of artificial neural networks. The optimum weights of artificial neural networks are determined in the training stage. The

performance of the proposed improved Nelder-Mead-Artificial neural networks classification algorithm has been tested on the most

common datasets from the UCI machine learning repository. The classification results obtained from the proposed improved Nelder-Mead-

Artificial neural networks classification algorithm are compared with the results of the standard Nelder-Mead-Artificial neural networks

classification algorithm. As a result of this comparison, the proposed improved Nelder-Mead-Artificial neural networks classification

algorithm has given best results in all datasets.

Keywords: Artificial neural networks, Nelder-Mead optimization method, Training algorithm.

1. Introduction

Artificial neural networks (ANNs) is a vital part of artificial

intelligence. Machine learning and cognitive sciences depend on

ANNs to solve various nonlinear mappings relationships [1].

ANN’s are normally used a back-propagation algorithm to fix

different problems on account of their approximation features.

With respecting to the weights back-propagation algorithm

computes explicit gradients of the error such as the mean of square

error. However, ANNs trained with gradient descent based

learning algorithm generally fall of slow convergence and local

minima. To get rid of these issues metaheuristic algorithm that uses

global search to find the solution has been used to train ANNs. For

instance, the genetic algorithm utilized to training ANNs [2];

another method employed in order to train artificial neural network

is reported by Slowik [3]. He used an adaptive differential

evolution algorithm to train ANNs. The aim of the calculation that

is achieved is to find optimal weights regarding an error rate [3].

Salman Mohaghegi et al. analyze the performance of particle

swarm optimization (PSO) and they compared its convergence

speed and robustness of the algorithm with the back-propagation

algorithm. The learning algorithms have been utilized in order to

modify the output synaptic weight array, and in each algorithm, the

centers and widths in the hidden layer are inferred utilizing an

offline clustering approach. According to the results, the PSO

algorithm showed good performance even with a fewnumbers of

particles. The statistical result proved PSO as a robust algorithm

for learning ANNs [4]. David J. Montana and Lawrence Davis

utilized an evolutionary algorithm for enhancing the performance

of feed word artificial neural networks. In addition, the algorithm

added more domain-specific knowledge into ANNs [5]. Malinak

and Jaksa utilized evolutionary algorithm with a local search

algorithm to get the best performance [6]. Kenndy and Eberhat

used swarm intelligence PSO for the first time [7]. Carvalho and

Ludermir applied PSO to medical benchmark problems. They

compared the statistical result with a local search operator in order

to learn ANNs. The result showed a better performance than other

well-studied algorithm [8]. Apart from PSO, the scientists applied

other swarm intelligence such as ant colony algorithm. Blum and

Socha employed a new version of ACO algorithm for training

artificial neural network [9]. Liang Hu, Longhui Qin, Kai Mao,

Wenyu Chen,and Xin Fu used a genetic algorithm for training

ANNs to solve the real-life problem of Multipath Ultrasonic Gas

Flowmeter. They used GANNs for multipath ultrasonic flowmeter

(UFM) to help decrease its error rate when detecting the flowrate

of the complicated flow field. The evaluation of results showed by

ANNs and GANNs demonstrated better achievement for ANNs

trained by genetic algorithm compared with once trained with

gradient descent based algorithm [10]. In addition to genetic

algorithm, to find the optimal weight for a neural network, the

genetic algorithm has been utilized to enhance the whole structure

for artificial neural networks. For example, K. G. Kapanova,

Dimov and M. Sellier utilized GA for optimizing the architecture

of ANNs. This approach allowed rearranging hidden layers,

neurons in every hidden layer, the connections between neurons

and the activation function [11].

In this study, we improved the Nelder-Mead optimization method

and used to determine the optimum weight values of ANNs.vIn the

proposed improved Nelder-Mead-ANNs classification algorithm,

ANNs classification algorithm performs well in the learning stage.

The paper is organized as follows: The ANNs classification

algorithm is explained in the second section. The standard Nelder-

Mead optimization method is explained in the third section. The

improved Nelder-Mead optimization method is explained in the

fourth section and the experimental results are discussed in the fifth

section. Finally, the conclusion is explained in the last section.

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Electronics and Computer Edu.,Selcuk University, Konya, Turkey 2 Stud. In computer Engg., Selcuk University, Konya, Turkey 3 Computer Engg.,Necmettin Erbakan University, Konya, Turkey

* Corresponding Author:Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 271–274 |272

2. Artificial Neural Networks

ANNs consist of neurons and these neurons that connect together.

These neurons can bind to each other very complexly as in the real

neurons system. Each neuron has a different weight input and one

output. For this purpose, the sum of the entries with different

weights is expressed by the following Equation 1 [12].

𝑛𝑒𝑡 = ∑ 𝐼𝑖𝑊𝑖𝑗

𝑚

𝑖=1

+ 𝑏 (1)

Where, m is the number of neurons in the layer, w is the weight

between i and j neuron and b is a bias neuron. The neuron output

value is obtained after the output of net function passes through the

activation function. Usually, in ANNs systems, the sigmoid

activation function is used. The sigmoid activation function is

shown in Equation 2 [13].

𝐹(𝑛𝑒𝑡) =1

1 + 𝑒−𝑛𝑒𝑡 (2)

The error rate between the output from the ANNs system and the

actual output is calculated. The mean square error is given in

Equation 3 below.

𝑀𝑆𝐸 =1

𝑛∑(𝑂𝑥 − 𝑇𝑥)2

𝑛

𝑥=1

(3)

Where, n representsa number of instances in datasets, Ox is the

output generated from the xth inputs and Tx is the target output of

the xth inputs. After training data is completed, the trained network

can estimate the result of any given dataset according to the last

state of the weight values. This process is called network training.

There are different training algorithms in network learning. In this

study, the Nelder-Mead optimization method is used in the

network training process.

3. Nelder-Mead optimization method

Nelder-Mead is a simple optimization method developed by

Nelder&Mead [17]. It is also called as Amoeba method in many

kinds of literature. It is a method widely used in multi-dimensional

unconstrained optimization problems. It is very simple to

understand the basic Nelder-Mead method and its usage is very

easy. For this reason, it is a very popular method especially in

many fields of chemistry and medicine and monolith. It is widely

used for solving parameter estimation and statistical problems. It

can also be used in the field of experimental mathematics [14].

The Nelder-mead optimization method is a simple method used to

find the local minimum of several variable functions generated by

Nelder and Mead. Simplex is an n-dimensional geometric shape. It

includes (N + 1) points in N dimensions. The simplex in two

dimensions consists of a triangle. It becomes a triangular prism in

three dimensions (4 surfaces). It is an example of a research

method that compares the function values at three corner points of

a triangle. By analyzing the function at all given points, the

program moves the highest and lowest values to the reflection

point relative to the surface above the highest and lowest elements.

In two dimensions, this method is a model search method that

compares the function values at the corners of a triangle. When we

consider the minimization state of a bivariate z=f (x, y) function,

the value of the z function at the worst corner of the triangle (w

worst vertex) is greatest.

For this reason, this worst corner is replaced by a new point. So we

get a new triangle. The search is now resumed in this new triangle,

so that the process becomes a triangle array where the function

value becomes smaller and smaller at the corner point value [15].

The steps of the Nelder-Mead optimization method are given

below.

𝑆𝑡𝑒𝑝 1: 𝐼𝑛𝑖𝑡𝑖𝑎𝑙𝑖𝑧𝑒 𝑤𝑖𝑡ℎ 𝑟𝑎𝑛𝑑𝑜𝑚 𝑠𝑜𝑙𝑢𝑡𝑖𝑜𝑛

𝑆𝑡𝑒𝑝 2: 𝑂𝑟𝑑𝑒𝑟 𝑠𝑜𝑙𝑢𝑡𝑖𝑜𝑛𝑠 𝑎𝑐𝑐𝑜𝑟𝑑𝑖𝑛𝑔 𝑡𝑜 𝑡ℎ𝑒 𝑓𝑖𝑡𝑛𝑒𝑠𝑠 𝑣𝑎𝑙𝑢𝑒𝑠(𝑥1) ≤ 𝑓(𝑥2) ≤ ⋯ ≤ 𝑓(𝑥𝑛+1)

𝑆𝑡𝑒𝑝 3: 𝐶𝑎𝑙𝑐𝑢𝑙𝑎𝑡𝑒 𝑥0 𝑡ℎ𝑒 𝑐𝑒𝑛𝑡𝑟𝑜𝑖𝑑 𝑜𝑓 𝑎𝑙𝑙 𝑝𝑜𝑖𝑛𝑡𝑠 𝑒𝑥𝑐𝑒𝑝𝑡 𝑥(𝑛+1)

𝑆𝑡𝑒𝑝 4: 𝐶𝑜𝑚𝑝𝑢𝑡𝑒 𝑟𝑒𝑓𝑙𝑒𝑐𝑡𝑖𝑜𝑛 𝑝𝑜𝑖𝑛𝑡 𝑥𝑟 = 𝑥0 + 𝛼(𝑥0 − 𝑥𝑛)

𝑆𝑡𝑒𝑝 5. 𝐼𝑓 (𝑥0 ) ≤ 𝑓(𝑥𝑟 ) < 𝑓(𝑥𝑛 ) 𝑡ℎ𝑒𝑛 𝑎𝑐𝑐𝑒𝑝𝑡 𝑥𝑟 𝑎𝑛𝑑 𝑡𝑒𝑟𝑚𝑖𝑛𝑎𝑡𝑒 𝑡ℎ𝑒 𝑖𝑡𝑒𝑟𝑎𝑡𝑖𝑜𝑛

𝑆𝑡𝑒𝑝 6. 𝐸𝑥𝑝𝑎𝑛𝑠𝑖𝑜𝑛, 𝐼𝑓 (𝑥𝑟) < 𝑓(𝑥1) 𝑡ℎ𝑒𝑛 𝑐𝑜𝑚𝑝𝑢𝑡𝑒 𝑡ℎ𝑒 𝑒𝑥𝑝𝑎𝑛𝑑𝑒𝑑 𝑝𝑜𝑖𝑛𝑡 𝑥𝑒 = 𝑥𝑟 + 𝛾(𝑥𝑟 − 𝑥0). 𝐼𝑓 (𝑥𝑒) < 𝑓(𝑥𝑟)

𝑎𝑐𝑐𝑒𝑝𝑡 𝑥𝑒𝑎𝑛𝑑 𝑡𝑒𝑟𝑚𝑖𝑛𝑎𝑡𝑒 𝑡ℎ𝑒 𝑖𝑡𝑒𝑟𝑎𝑡𝑖𝑜𝑛, 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒 𝑎𝑐𝑐𝑒𝑝𝑡 𝑥𝑟𝑎𝑛𝑑 𝑡𝑒𝑟𝑚𝑖𝑛𝑎𝑡𝑒 𝑡ℎ𝑒 𝑖𝑡𝑒𝑟𝑎𝑡𝑖𝑜𝑛. 𝑆𝑡𝑒𝑝 7. 𝐶𝑜𝑛𝑡𝑟𝑎𝑐𝑡𝑖𝑜𝑛, 𝐼𝑓 (𝑥𝑟) ≥ 𝑓(𝑥𝑛) 𝑝𝑒𝑟𝑓𝑜𝑟𝑚 𝑎 𝑐𝑜𝑛𝑡𝑟𝑎𝑐𝑡𝑖𝑜𝑛 𝑏𝑒𝑡𝑤𝑒𝑒𝑛 𝑥0𝑎𝑛𝑑 𝑥𝑛. 𝑥𝑐 = 𝑥0 + 𝜌(𝑥0 − 𝑥𝑛). 𝐼𝑓 (𝑥𝑐) < 𝑓(𝑥𝑛) 𝑎𝑐𝑐𝑒𝑝𝑡 𝑥𝑐 𝑎𝑛𝑑 𝑡𝑒𝑟𝑚𝑖𝑛𝑎𝑡𝑒 𝑖𝑡𝑒𝑟𝑎𝑡𝑖𝑜𝑛. 𝑆𝑡𝑒𝑝 8. 𝑆ℎ𝑟𝑖𝑛𝑘, 𝑅𝑒𝑝𝑙𝑎𝑐𝑒 𝑎𝑙𝑙 𝑝𝑜𝑖𝑛𝑡𝑠 𝑒𝑥𝑐𝑒𝑝𝑡 𝑡ℎ𝑒 𝑏𝑒𝑠𝑡 𝑥1

𝑤𝑖𝑡ℎ 𝑥𝑖 = 𝑥0 + 𝜎(𝑥𝑖 − 𝑥1) 𝑎𝑛𝑑 𝑔𝑜 𝑡𝑜 𝑠𝑡𝑒𝑝 2.

4. Improved Nelder-Mead optimization method

Since the Nelder-Mead optimization method is simple, it is also

known as a simple optimization method. In this study, the Nelder-

Mead optimization method is combined with PSO algorithm to

find optimum value of ANNs weights. In the proposed

classification algorithm, firstly, the weights of the ANNs

classification algorithm are generated randomly and the MSE is

calculated, then the best result is selected. The weights are updated

according to the best solution obtained from random solutions

according to the Nelder-Mead optimization method. Then the best

MSE is selected. In the next iteration, if the difference error of the

previous iteration and of the current error is less than 0.5 update

the weights according to the velocity equation of the PSO

algorithm is given in equations 4 and 5 below.

𝑣𝑖𝑡+1 = 𝑤 ∗ 𝑣𝑖

𝑡 + 𝑐1𝑟1(𝑃𝑏𝑒𝑠𝑡𝑖 − 𝑥𝑖𝑡)

+ 𝑐2𝑟2(𝐺𝑏𝑒𝑠𝑡 − 𝑥𝑖𝑡)

(4)

𝑥𝑖𝑡+1 = 𝑥𝑖

𝑡 + 𝑣𝑖𝑡+1 (5)

Where, t is the number of iteration, w is inertia weight, d is

dimension, xid current position of particle, vt the previous velocity

of particle, Pbesti is a better fitness value, Gbest is a better fitness

value of all particles, c1 and c2 are two positive constants represents

accelerate factor and r1 and r2 are two random functions in the

range [0,1]. Otherwise, the weight updating according to the

Nelder-Mead optimization method functions. Optimization

algorithms are used in many engineering areas and aim to find the

best end result. In this study, the improved Nelder-Mead

optimization method is used to find the optimum weights in the

ANNs classification model. The flowchart of the proposed ANNs

classification model is given in Fig. 1 below.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 271–274 |273

Fig. 1. Flowchart of the proposed ANNs model

5. The experimental results

In the experimental results section, the performance of the two

methods is compared by applying the developed Nelder-Mead

method and the standard Nelder-Mead method in the training of

the ANNs classification algorithm. The codes of standard Nelder-

Mead and developed Nelder-Mead methods were written in the

visual studio. The performance of the classification algorithms was

applied to nine UCI datasets [16]. The characteristic of datasets is

given in Table 1.

Table 1. The characteristic of datasets

Datasets name No. of

Attributes

No. of

sample

No. of

classes

Balance 4 625 3 Shuttle-Landing 6 253 2

lymphography 18 148 4

Breast cancer 9 699 2 Iris 4 150 3

Wine 13 178 3

Thyroid 5 215 3 E.coli 7 336 8

Glass 9 214 6

An appropriate ANNs was created for each dataset and the number

of nodes in the hidden layer was set. Bias node was used in the

hidden and output layers and sigmoid was used as the activation

function. In the Nelder-Mead optimization method, the values α, γ,

ρ, α, and σ were set to 2, 0.5, 1, and 0.5, respectively. Initial values

of weights were between 10 and -10 and the algorithm was run at

1000 iterations. MSE was used as an activation function in both

the Nelder-Mead optimization method. The classification

performance of the algorithm is calculated based on the final value

of this fitness function. 10-Fold cross-validation was used for

classification accuracy (CA).

𝐶𝐴 =𝑁𝑜. 𝑜𝑓 𝑐𝑜𝑟𝑟𝑒𝑐𝑡 𝑝𝑟𝑒𝑑𝑖𝑐𝑡𝑒𝑑 𝑠𝑎𝑚𝑝𝑙𝑒𝑠

𝑁𝑜. 𝑜𝑓 𝑡𝑜𝑡𝑎𝑙 𝑠𝑎𝑚𝑝𝑙𝑒𝑠 (6)

The classification results of the proposed improved Nelder-Mead-

ANNs classification algorithm and the standard Nelder-Mead-

ANNs classification algorithm is given in table 2 below. In this

table, the CA and MSE obtained in the training and testing stages

of both classification algorithms are given. Also, H is the optimal

number of neurons in a hidden layer.

When we examine the experimental results of Table 2, the

proposed Nelder-Mead-ANNs classification algorithm showed

better classification accuracy and MSE in alldatasets than the

standard Nelder-Mead-ANNs classification algorithm.

6. Conclusion

In our study, we performed the training of the ANNs classification

algorithm with standard Nelder-Mead and improved Nelder-Mead

optimization methods. The improved Nelder-Mead optimization

method was developed because the standard Nelder-Mead

optimization method did not fulfill this task properly. With the help

of the PSO optimization method, the developed Nelder-mead

optimization method gave better results in the ANNs training. The

improved Nelder-Mead-ANNs classification algorithm was

applied to 9 datasets of UCI machine learning repository and

performed better on all datasets. So the developed classification

Table 2. The classification results of the Standard Nelder-Mead-ANNs and Improved Nelder-Mead-ANNsclassification algorithms

Datasets name Standard Nelder-Mead-ANNs Improved Nelder-Mead-ANNs

Training Testing MSE H Training Testing MSE H

Balance 0,789 0,792 0,330 9 0,908 0,902 0,138 3

Shuttle-Landing 0,889 0,892 0,181 7 0,990 0,955 0,015 7 lymphography 0,615 0,648 0,515 12 0,902 0,710 0,169 22

Breast cancer 0,970 0,970 0,055 10 0,983 0,990 0,031 5

Iris 0,845 0,892 0,291 9 0,988 0,953 0,024 8 Wine 0,992 0,928 0,42 12 0,984 0,931 0,030 17

Thyroid 0,816 0,802 0,325 3 0,983 0,956 0,029 10

E.coli 0,592 0,577 0,602 4 0,840 0,810 0,276 10 Glass 0,375 0,386 0,748 4 0,682 0,619 0,507 11

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 271–274 |274

algorithm can be applied to many fields such as medical,

engineering and recognition systems.

Acknowledgement

This study presents the results of Master thesis of Mustafa Merdan in

Selcuk University Natural and Applied Science Institute Computer

Engineering Programme.

References

[1] APA Kecman, V. (2001). Learning and soft computing: support

vector machines, neural networks, and fuzzy logic models. MIT press.

[2] Alhamdoosh, M., & Wang, D. (2014). Fast decorrelated neural

network ensembles with random weights. Information Sciences, 264,

104-117.

[3] Slowik, A. (2011). Application of an adaptive differential evolution

algorithm with multiple trial vectors to artificial neural network

training. IEEE Transactions on Industrial Electronics, 58(8), 3160-

3167.

[4] Mohaghegi, S., del Valle, Y., Venayagamoorthy, G. K., & Harley, R.

G. (2005, June). A comparison of PSO and backpropagation for

training RBF neural networks for identification of a power system with

STATCOM. In Swarm Intelligence Symposium, 2005. SIS 2005.

Proceedings 2005 IEEE (pp. 381-384).

[5] Montana, D. J., & Davis, L. (1989, August). Training Feedforward

Neural Networks Using Genetic Algorithms. In IJCAI (Vol. 89, pp.

762-767).

[6] Malinak, P., &Jaksa, R. (2007, September). Simultaneous gradient and

evolutionary neural network weights adaptation methods. In

Evolutionary Computation, 2007. CEC 2007. IEEE Congress on (pp.

2665-2671). IEEE.

[7] Kennedy, J. (2011). Particle swarm optimization. In Encyclopedia of

machine learning (pp. 760-766). Springer US.

[8] Carvalho, M., &Ludermir, T. B. (2007, September). Particle swarm

optimization of neural network architectures and weights. In Hybrid

Intelligent Systems, 2007. HIS 2007. 7th International Conference on

(pp. 336-339). IEEE.

[9] Dorigo, M., Birattari, M., &Stutzle, T. (2006). Ant colony

optimization. IEEE computational intelligence magazine, 1(4), 28-39.

[10] Hu, Liang, et al. "Optimization of neural network by genetic algorithm

for flowrate determination in multipath ultrasonic gas flowmeter."

IEEE Sensors Journal 16.5 (2016): 1158-1167.

[11] Kapanova, K. G., Dimov, I., &Sellier, J. M. A genetic approach to

automatic neural network architecture optimization. Neural

Computing and Applications, 1-12.

[12] Han, J., Pei, J., &Kamber, M. (2011). Data mining: concepts and

techniques. Elsevier.

[13] Schalkoff, R. J. (1997). Artificial neural networks (Vol. 1). New York:

McGraw-Hill.

[14] Kelley, C. T. (1999). Detection and remediation of stagnation in the

nelder-mead algorithm using a sufficient decrease condition. SIAM

journal on optimization, 10(1), 43-55.

[15] Lagarias J. C., J. A. Reeds, M. H. Wright, and P. E. Wright, 1998.

Convergence properties of the Nelder-Mead simplex method in low

dimensions. SIAM Journal on Optimization, 9(1), p. 112- 147.

[16] Asuncion, A., & Newman, D. (2007). UCI machine learning

repository.

[17] Nelder, John A., R. Mead (1965). A simplex method for function

minimization. Computer Journal. 7: 308–313.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 275–281 | 275

A fuzzy-genetic based design of permanent magnet synchronous motor

Mumtaz MUTLUER1*

Accepted : 08/11/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: This paper presents a fuzzy-genetic based design of permanent magnet synchronous motor. The selected motor structure with

surface magnet and double layer winding is for high torque and low speed applications. The design approach involves combining fuzzy

logic and genetic algorithm in a powerful combination. While the genetic algorithm is used in scanning of the solution space, the fuzzy

logic approach has been utilized in selecting the most appropriate solutions. While choosing geometric parameters as input for optimization,

design equations are obtained by using geometrical, electrical and magnetic properties of the motor. The output results are evaluated with

motor efficiency, motor weight and weight of magnets as the objective function. Furthermore, the multiobjective design optimization

results are compared with the results obtained for each single objective and tested with a finite element program. The results are finally

remarkable and quite compatible with the finite element results.

Keywords: Design optimization, fuzzy-genetic approach, permanent magnet synchronous motor

1. Introduction

Induction motor and permanent magnet synchronous motor

(PMSM) are among the most used electric motors in industrial

fields. While induction motors are of interest because of their low

cost and ease of maintenance, permanent magnet synchronous

motors have high power density and high efficiency. In addition,

in today’s control applications, drive systems affect the motor

selection. After all, whatever the performance criteria, due to high

efficiency it is clear that the use of PMSMs is increasing. Most

prominent feature of PMSMs is that they show structural

differences according to the placement of the magnets. Naturally

this affects the performances and the production costs of PMSMs.

Due to the ease of design and low production cost, surface mounted

PMSMs are the most preferred types for low speed applications.

This structure is also preferred because of its low cogging torque

based on slot and pole combination [1].

The design of electric motors is the most complex engineering

problem. It is because the electric motors have the non-linear

structures. To overcome this situation, when some simplifications

are taken, linear equations are used in electric motor design studies.

This is even more preliminary in optimization applications wherein

artificial intelligence algorithms are used. In academic or industrial

fields complex and realistic designs are provided by commercial

programs using finite element method. In fact, analytically and

numerically, two methods are basically applied in the design of

electric motors. Working with the analytic equations is weak in

terms of accuracy but advantageous in terms of duration. On the

other hand working with the numerical methods such as finite

element method is effective in terms of correctness of the results

but weak in terms of duration [2]. Nevertheless, in the design

optimization studies where analytic equations are used, electric

motor designs can be flexibly directed to single objective or

multiobjective.

With a general evaluation, the design optimization studies which

artificial intelligence used are based on the effectiveness of the

algorithms. One of them is undoubtedly the most recognizable

genetic algorithm. This algorithm has been applied to so many

different electric motor design problems up to date [3-5]. In

addition, the more strong algorithms constructed by combining the

different artificial intelligence algorithms or other methods with

genetic algorithm have been used [6-9].

The designs of permanent magnet synchronous motors are more

complex and non-linear engineering problems and also contain

some fuzzy facts such as other electric motor designs [10, 11]. The

fuzzy-genetic approach is quite useful in terms of decision

structure and ease of application in the choice of objective

function. By using the fuzzy-genetic structure, a large solution

space can be scanned and the designer’s experience, view and

judgment are well reflected [10-14]. In this study, as

multiobjective, a fuzzy-genetic based approach was firstly used in

the design optimization of the surface mounted permanent magnet

synchronous motor. Motor efficiency, motor weight and weight of

magnets were selected as objective function and then the objective

and constraint values were determined by using fuzzy rules. Single

and multi objective optimization results were given comparatively,

and the results were tested by a finite element program. The effects

of the fuzzy-genetic structure used were shown in the PMSM

design with graphics and tables. In this respect, PMSM design

optimization has been examined in a versatile way and useful

inferences have been provided for the technical staffs.

2. Multidirectional Analysis of The PMSM

Electrical, magnetic, thermal and mechanical analyzes are carried

out in the designs of electric motors. The solution of the differential

equations obtained for each analysis is quite complex. Moreover,

due to the non-linearity of the electric motors, it is almost

impossible to do very precise solutions. Numerical methods such

as finite element method are used in this case. Obviously, the use

of linear equations for a basic design is sufficient. The geometrical

model of the PMSM, the electrical and the magnetic circuits are

investigated in each subdivision as follows.

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

_

1Department of Electrical and Electronic Engineering, Necmettin Erbakan

University, Konya, Turkey

* Corresponding Author Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 275–281 | 276

2.1. The Geometric Model

The use of magnetic and electrical equations commonly associated

with the geometric modelling in the designs of electric motors is

widespread. Such approaches are primitive but provide rapid

analysis [2]. In this study based on the geometric model of the

PMSM has 12 slots and 10 poles shown in Figure 1. By some

assumptions, the objective functions are obtained with the help of

magnetic and electric equations. Facilitating acceptance is the

result of an optimal design that sets the grounds for these

assumptions.

Figure 1. 3D geometric model of the PMSM

2.2. The Magnetic Circuit

According to the magnetic circuit in Figure 2 [1, 2] it is very

difficult to calculate the magnetic flux in each region of the

PMSM. The most important issue in the magnetic design is the

accurate calculation of air gap magnetic flux [15]. The magnetic

flux at the points on the stator and the rotor is particularly

important in terms of saturation. From this point of view, magnetic

boundary values are the points to be considered in the PMSM

design. Magnetic flux density equations of air gap, stator tooth,

stator yoke and rotor yoke are as follows.

𝐵𝑚 = (𝐵𝑟𝑘𝑙𝑒𝑎𝑘𝑙m) (𝑙m + 𝜇𝑟𝛿𝑘𝐶)⁄ (1)

�̂�𝛿 = (4 π⁄ )𝐵𝑚 sin α (2)

𝐵𝑠𝑡 = (4𝛼𝐵𝑚(𝐷 2⁄ − 𝛿)(1 − 𝑘𝑙𝑒𝑎𝑘𝑡𝑜𝑜𝑡ℎ)) (2𝑝𝑞𝑏𝑠𝑡𝑠𝑡𝑓𝑐)⁄ (3)

𝐵𝑠𝑦 = (4𝛼𝐵𝑚(𝐷 2⁄ − 𝛿)) (2𝑝ℎ𝑠𝑦𝑠𝑡𝑓𝑐)⁄ (4)

𝐵𝑟𝑦 = (4𝛼𝐵𝑚(𝐷 2⁄ − 𝛿)) (2𝑝ℎ𝑟𝑦𝑠𝑡𝑓𝑐)⁄ (5)

where, remanence flux density of permanent magnet is 𝐵𝑟,

maximum of air gap flux density is 𝐵𝑚, fundamental of air gap flux

density is �̂�𝛿 , flux density in a stator tooth is 𝐵𝑠𝑡, flux density in

stator yoke is 𝐵𝑠𝑦 , flux density in rotor yoke is 𝐵𝑟𝑦, correction

factor for air gap flux density is 𝑘𝑙𝑒𝑎𝑘, relative magnet permeability

is 𝜇𝑟 , Carter factor is 𝑘𝐶 , pole angle is 2𝛼, inner stator diameter is

𝐷, correction factor for flux density in stator teeth is 𝑘𝑙𝑒𝑎𝑘𝑡𝑜𝑜𝑡ℎ,

number of slots per pole per phase is 𝑞, stacking factor of the stator

iron laminations is 𝑠𝑡𝑓𝑐, stator yoke height is ℎ𝑠𝑦, rotor yoke

height is ℎ𝑟𝑦.

Figure 2. Magnetic circuit for the geometrical model of the PMSM

2.3. The Electrical Circuits

Electromechanical conversions are the most important part of

electrical circuit design. Here, the motor d-q electrical circuits at

base speed (Fig. 3) and the equations were given. When the

moment equation is examined, the number of windings of the

motor is calculated only according to the 𝐼𝑞 current because the 𝐼𝑑

current is zero in the non-salient permanent magnet synchronous

motors [16, 17].

�̂� = 𝜔𝑘𝜔1𝑞𝑛𝑠�̂�𝛿𝐿(𝐷 − 𝛿) (6)

𝑅𝐶𝑢 = (𝜌𝐶𝑢(𝑝𝐿 + 𝜋𝑘𝑐𝑜𝑖𝑙(𝐷 + ℎ𝑠𝑠))𝑛𝑠2𝑞) 𝑓𝑠𝐴𝑠𝑙⁄ (7)

𝐿𝑑,𝑞 = (𝑝𝑞𝜆 + 3 𝜋⁄ (𝑞𝑘𝜔1)2 (𝐷 − 𝛿) (𝛿𝑘𝐶 + 𝑙𝑚 𝜇𝑟⁄ )⁄ )𝜇0𝐿𝑛𝑠

2(8)

�̂� = √𝑈𝑞2 +𝑈𝑑

2 = √(�̂� + 𝑅𝐶𝑢𝐼𝑞)2+ (𝐿𝑞𝜔𝐼𝑞)

2 (9)

where, electrical angular frequency is 𝜔, fundamental winding

factor is𝑘𝜔1, conductor number per slot is 𝑛𝑠, stack length is 𝐿,

copper wire resistivity is 𝜌𝐶𝑢, end-winding coefficient is 𝑘𝑐𝑜𝑖𝑙 , slot

fill factor is 𝑓𝑠, slot area is 𝐴𝑠𝑙, specific permeance coefficient of

the slot opening is 𝜆, d,q-axes terminal voltages are is 𝑈𝑑,𝑞, d,q-

axes currents are 𝐼𝑑,𝑞 , fundamental of the induced voltage is �̂�,

winding resistance is 𝑅𝐶𝑢, d,q-axes magnetizing inductance is 𝐿𝑑,𝑞.

Figure 3. d-q equivalent circuits of the PMSM

2.4. The Objective Functions

The efficiency is generally the primary objective in the design of

the electrical motor today. To improve efficiency in design studies,

the reduction of copper losses is especially required for low-

frequency multi-poles PMSMs. It is also necessary to pay more

attention to one issue that the gearless PMSMs are more efficient

than other electric motors with gears.

Motor weight and weight of magnets affect the cost as much as it

is important in terms of the usage place. The permanent magnets

are structurally the most expensive parts of the PMSMs, and their

IqRCu

ωLdId

EUq

IdRCu

Ud ωLqIq

q-axis equivalent circuit d-axis equivalent circuit

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 275–281 | 277

prices are also rapidly changing, especially due to technological

developments. The weight of magnets improves the power density

of the PMSMs and but increases the motor weight. Especially this

situation is overwhelmed by multi-poles PMSM structures.

Three objective functions, namely motor efficiency, motor weight

and weight of magnets, were used for the fuzzy-genetic based

multiobjective design optimization. A suitable association for all

objectives was aimed, namely to increase the motor efficiency, to

reduce the motor weight and to reduce the weight of the magnets.

The acquisition of these functions is a very detailed process and

therefore references to different studies have been made [1, 2, 16-

19]. As a result, the objective functions required for this study were

obtained in the following order, motor efficiency, motor weight

and weight of magnets.

𝜂 = 𝑃𝑜𝑢𝑡 (𝑃𝑜𝑢𝑡 + 𝑃𝐶𝑢 + 𝑃𝐹𝑒)⁄ (10)

𝑊𝑇𝑜𝑡 = 𝑊𝑆ℎ𝑎𝑓𝑡 +𝑊𝑃𝑀𝑠 +𝑊𝑅𝑜𝑡𝑜𝑟 +𝑊𝑆𝑡𝑎𝑡𝑜𝑟 +𝑊𝑊𝑖𝑛𝑑𝑖𝑛𝑔 (11)

𝑊𝑃𝑀 = 𝜌𝑃𝑀4𝛼𝐿𝑙𝑚(𝐷𝑟𝑐 + 𝑙𝑚) (12)

where, efficiency is 𝜂, output power is 𝑃𝑜𝑢𝑡, copper losses is 𝑃𝐶𝑢,

iron losses are 𝑃𝐹𝑒, total motor weight is 𝑊𝑇𝑜𝑡, shaft weight is

𝑊𝑆ℎ𝑎𝑓𝑡, rotor weight is 𝑊𝑅𝑜𝑡𝑜𝑟, stator weight is 𝑊𝑆𝑡𝑎𝑡𝑜𝑟, winding

weight is 𝑊𝑊𝑖𝑛𝑑𝑖𝑛𝑔, weight of magnets is 𝑊𝑃𝑀𝑠.

3. The Fuzzy-Genetic Based Multiobjective Approach

The genetic algorithm has been applied to so many different

electric motor design problems up to date [3-5]. In addition, the

more strong algorithms constructed by combining the different

artificial intelligence algorithms or other methods with genetic

algorithm have been used [6-9]. In addition, the fuzzy-genetic

approach has been used to solve some engineering problems such

as selection of control parameters and induction motor design and

so effective solutions have been made. Here, the definition and

basic steps of the used fuzzy-genetic based multiobjective

approach with the genetic algorithm are as follows [20, 21]:

𝑓𝑖𝑛𝑑𝑋𝑤ℎ𝑖𝑐ℎ𝑚𝑎𝑥𝑖𝑚𝑖𝑧𝑒 𝑚𝑖𝑛𝑖𝑚𝑖𝑧𝑒⁄ 𝑓(𝑋)

𝑠𝑢𝑏𝑗𝑒𝑐𝑡𝑡𝑜; ℎ𝑚(𝑋) = 0, 𝑚 = 1,2,… ,𝑀

𝑔𝑗(𝑋) ≤ 0, 𝑗 = 1,2,… , 𝐽

𝑋𝑘𝑙 ≤ 𝑋𝑘 ≤ 𝑋𝑘

𝑢, 𝑘 = 1,2,… , 𝐾

where objective functions form the multiobjective function vector

as 𝑓(𝑥) = [𝑓1(𝑋), 𝑓2(𝑋), . . . , 𝑓𝑛(𝑋)], ℎ𝑚(𝑋) and 𝑔𝑗(𝑋) are

equality and inequality constraint functions. 𝑋𝑘𝑢 and 𝑋𝑘

𝑙 are the

upper and lower boundary values of the input parameter. The fuzzy

objective is obtained by using the actual objective function values

within the fuzzy boundaries. It uses the fuzzy membership s-

function defined below.

𝜇𝑓𝑖(𝑋) =

{

0

1 − 2 (𝑓𝑖(𝑋)−𝑓𝑖

𝑚𝑖𝑛

𝑓𝑖𝑚𝑎𝑥−𝑓𝑖

𝑚𝑖𝑛)2

2 (𝑓𝑖𝑚𝑎𝑥−𝑓𝑖(𝑋)

𝑓𝑖𝑚𝑎𝑥−𝑓𝑖

𝑚𝑖𝑛)2

1

𝑖𝑓 𝑓𝑖(𝑋) ≤ 𝑓𝑖𝑚𝑖𝑛

𝑖𝑓 𝑓𝑖𝑚𝑖𝑛 < 𝑓𝑖(𝑋) ≤ (𝑓𝑖

𝑚𝑎𝑥 + 𝑓𝑖𝑚𝑖𝑛) 2⁄

𝑖𝑓 (𝑓𝑖𝑚𝑎𝑥 + 𝑓𝑖

𝑚𝑖𝑛) 2⁄ < 𝑓𝑖(𝑋) ≤ 𝑓𝑖𝑚𝑎𝑥

𝑖𝑓 𝑓𝑖𝑚𝑎𝑥 ≤ 𝑓𝑖(𝑋)

(13)

where 𝜇𝑓𝑖(𝑋):ℛ𝑛 → [0,1] and it represents fuzzy correctness

according to input parameters, 𝑓𝑖𝑚𝑎𝑥 and 𝑓𝑖

𝑚𝑎𝑥 expressions are

user-dependent minimum and maximum objective values. The

fuzzy values of the objective function and the constraint values for

fuzzy decision making must be calculated, and the conclusion for

the constraint function is as follows.

𝜇𝑔𝑗(𝑋) = {0

1 − (𝑔𝑗(𝑋) − 𝑏𝑗) 𝑑𝑗⁄

1

𝑖𝑓 𝑔𝑗(𝑋) > 𝑏𝑗 + 𝑑𝑗𝑖𝑓 𝑏𝑗 ≤ 𝑔𝑗(𝑋) ≤ 𝑏𝑗 + 𝑑𝑗

𝑖𝑓 𝑔𝑗(𝑋) < 𝑏𝑗

(14)

where 𝜇𝑔𝑗(𝑋):ℛ𝑛 → [0,1] and it represents fuzzy constraints

according to input parameters. 𝑏𝑗 is the desired value, 𝑑𝑗 is the

tolerance value.

3.1. Fuzzy Decision Making

The aim of the multiobjective optimization is to find the best

solution by using linearized fuzzy objective and constraint

membership functions. That is to find a solution with maximum

membership from the fuzzy solution space. This situation can be

shown as follows [20, 22]:

𝜇𝐷(𝑋∗) = max(𝜇𝐷(𝑋)) , 𝜇𝐷 ∈ [0,1] (15)

Convex fuzzy decision criterion was chosen in this study. Convex

decision is an approach that depends on the arithmetic mean and

the weight of each fuzzy objective function. As shown below.

𝐷 = 𝛼𝑓(𝑋) + 𝛽𝑔(𝑋) (16)

where 𝛼 and 𝛽 are weighting factors, which satisfy

𝛼 + 𝛽 = 1 (𝛼 ≥ 0 ; 𝛽 ≥ 0) (17)

These coefficients can be obtained by the linear weight average of

the objective functions. Thus, the membership function is provided

for convex fuzzy inference.

𝛼𝑖 = 𝜇𝑓𝑖 ∑ 𝜇𝑓𝑖𝑛𝑖=1⁄ 𝑎𝑛𝑑𝛽𝑗 = 𝜇𝑔𝑗 ∑ 𝜇𝑔𝑗

𝑚𝑗=1⁄ (18)

𝜇𝐷(𝑋) = ∑ 𝛼𝑖𝜇𝑓𝑖𝑛𝑖=1 + ∑ 𝛽𝑗𝜇𝑔𝑗

𝑚𝑗=1 (19)

where 𝛼𝑖 and 𝛽𝑗 satisfy

∑ 𝛼𝑖𝑛𝑖=1 + ∑ 𝛽𝑗

𝑚𝑗=1 = 1

𝛼𝑖 ≥ 0 𝑖 = 1,2, … , 𝑛𝛽𝑗 ≥ 0 𝑗 = 1,2, … ,𝑚

(20)

The multiobjective design algorithm used here includes properties

of fuzzy logic and genetic algorithm. While fuzzy logic approach

reflects human thought in selecting the best result of the solution

spaces, genetic algorithm tries to find the most appropriate result

in a large solution space. Hence, the hybrid algorithm will strongly

reflect the goals of the motor design.

The starting point is the geometric parameters in multiobjective

design optimization. Selected geometric parameters are

independent variable. Supply voltage, motor power and speed, etc.

the quantities are invariable. Other design parameters are

dependent variable. The objective functions were obtained by

using the geometrical model, electrical and magnetic circuits of the

motor. The most important aspect of optimization studies is the

necessity of achieving objective functions with great accuracy.

Objective functions for fuzzy multiobjective design optimization

are formulated in different forms and compared with single

objective studies with genetic algorithm. These formulations are

shown in Table 1.

Table 1. Objective functions for the design optimization

Single objective functions Multiobjective functions

𝑓1(to increase motor efficiency) 𝐹1: 𝑚𝑎𝑥[𝜇𝑓1, 𝜇𝑓2]

𝑓2(to reduce motor weight) 𝐹2: 𝑚𝑎𝑥[𝜇𝑓1, 𝜇𝑓3]

𝑓3(to reduce magnet weight) 𝐹3:𝑚𝑎𝑥[𝜇𝑓1 , 𝜇𝑓2, 𝜇𝑓3]

The boundary values for fuzzification of the selected objective

functions are given in Table 2. The appropriate values and

tolerances for fuzzification of the constraints are given in Table 3.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 275–281 | 278

Table 2. Boundary values of the objective functions for fuzzification

Objective Functions Min value Max value

𝑓1(%) 0 100

𝑓2(𝑘𝑔) 50 150

𝑓3(𝑘𝑔) 0 3

Table 3.Constraint values and tolerances for fuzzification

Constraint Desired value

(𝒃𝒋) Tolerance value

(𝒅𝒋)

Stator tooth flux (𝐵𝑠𝑡) 1.6T 0.4T

Stator yoke flux

(𝐵𝑠𝑦) 1.4T 0.4T

Rotor yoke flux (𝐵𝑟𝑦) 1.4T 0.4T

3.2. Steps of the optimization algorithm

The multiobjective function can be written as Eq. 21 and the

flowchart of the optimization algorithm and the content of each

step were explained in detail.

𝑚𝑎𝑥(𝜇𝐷𝑖) = ∑ 𝛼𝑖𝜇𝑓𝑖𝑛𝑖=1 + ∑ 𝛽𝑗𝜇𝑔𝑗

𝑚𝑗=1 (21)

It is not important to produce the first population, because the

performance of artificial intelligence algorithms such as genetic

algorithm is not dependent on initial population or individuals.

However, the performance of the algorithm is affected by real-

valued or binary coding [23], where the binary code was used for

software ease. Also, since electric motor design studies require a

lot of equations, the boundaries of the independent variables must

be carefully chosen according to experience and requirements. The

geometrical, electrical and magnetic equations are interaction with

each other. It is therefore an effective approach to derive design

equations by making use of design experience to make some

negligence or to reduce motor design equations by using

coefficients. Thus, an effective and useful approach for the

designer will be achieved.

The obtained objective values are evaluated according to the

previously described fuzzy decision approach. In this way, new

populations are provided for the objective of the genetic algorithm

and higher values are preserved. The termination criterion for the

genetic algorithm was given as the iteration number or it could be

the precision of the objective values. The important point here is to

achieve the objectives with great accuracy.

The characteristics of the genetic algorithm and how it works are

as follows [23-25].

i. It works to select the best individual (solution) and more

solution is produced with the population for the optimization

problem. Individuals in the population are independent of

each other, whereas individuals are made up of genes that

contain the solution of each independent variable. The

population size and the number of genes are related to the

direct input parameters, which influences the solution

accuracy [17].

ii. Genetic algorithms carry the genetic properties of

individuals to new populations by using the fitness function.

Due to natural selection, strong individuals are more

fortunate to survive from weaker individuals. This situation

is repeated in each iteration to converge to the optimal

solution.

iii. Genetic algorithms do not guarantee the optimal solution.

Unfortunately, genetic algorithms can converge to a local

solution. Genetic algorithms have reproduction, crossover

and mutation operators. Using roulette wheel, tournament, or

different crossover operators, individuals with high fitness

values are selected. The crossover operator randomly

changes the genes of two selected individuals. The mutation

operator changes the gene of the preselected individual to "0"

or vice versa for binary coding. In this way, the algorithm is

prevented from converging with local solutions.

4. The Design Application and The Evaluation of The Results

Some constants must be predetermined in the design optimization.

In this study permanent magnet synchronous motor with

concentrated double layer winding has 340 volts of supply voltage,

2400 watts of shaft power, 250 rpm and also outer stator diameter

is 300mm, stack length is 120mm and electrical magnet angle is

126°. After the invariables determined, variables and their

boundary values was given in Table 4. The variable number was

chosen to be sufficient for the geometry of the PMSM.

Incorporating too many variables into the algorithm will affect

sensitivity of the results and the optimization time.

Table 4.Input variables of the design optimization

Parameter Symbol Lower

boundary

Upper

boundary

Magnet thickness (mm) 𝑙𝑚 3.5 5.5

Air gap length (mm) 𝛿 1 1.5

Slot wedge height (mm) ℎ𝑠𝑤 0.5 3

Stator tooth width (mm) 𝑏𝑡𝑠 10 50

Outer rotor diameter (mm) 𝐷𝑟𝑐 150 250

Stator slot height (mm) ℎ𝑠𝑠 25 45

Ratio of the slot opening over the slot width

𝑘𝑜𝑝𝑒𝑛 0.55 0.99

No initial solution is predicted in the design optimization study

made with single and multi objectives. The optimization results

were randomly generated by algorithms using the objective

function. In the case of single objective operation, the motor

efficiency, the total motor weight, the total weight of the magnets

have been tried to be improved separately and the result graphs are

shown in Figs. 4 and 5. In the case of multiobjective operation, the

motor efficiency, the total motor weight, the total weight of the

magnets have been tried to be mixed with fuzzy decision making

and the membership result graphs are given in Fig. 6. Algorithm

parameters such as population number are 100, iteration number is

100, crossover rate is 0.85 and mutual rate is 0.01 are taken the

same in order to evaluate the optimization results correctly. When

these graphs are examined, it can be said that genetic algorithm

provides a good search and solution in single objective

optimizations. In multiobjective optimizations, it can be stated that

the fuzzy decision making and genetic algorithm provide a good

fit and precise convergence. Furthermore, the selected population,

iteration, crossover and mutation values are suitable.

Table 5 shows the optimal geometric results and Table 6 shows the

optimal objective results, the optimal boundary results and the

iteration numbers and times in which the optimal results were

obtained. The values given in bold are the most appropriate results

found. According to Table 5, a common observation in all

iterations is the reduction of the rotor diameter, the increase of the

magnet thickness and the increase of the stator slot height. The

rotor diameter and slot height are interconnected. The PMSM

needs sufficient magnet weight and winding area to provide the

required torque. Table 6 emphasizes that results are obtained in

accordance with each objective function and the boundary values

are not exceeded. The situation that affects the iteration numbers

and times is in particular the effectiveness of the genetic algorithm.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 275–281 | 279

Figure 4.Optimal values of f1 single objective function

Figure 5.Optimal values of f2and f3 single objective functions

Figure 6.Convergence graphics of F1, F2 and F3 multiobjectivefunctions

Figure 7.Sum of the scaled objective results

Fig. 7 shows linear curves showing the most appropriate solution

for each row in Table 3. Each curve connects the points

corresponding to sum of the scaled values of the three accepted

objectives. According to Fig. 7, the motor efficiency and the motor

weight are not changed sharply in the optimization, but the weight

of the magnets sharply changes and affects the results. Linear

curves emphasize that the weight of the magnets is a very effective

objective function for the boundary values.

Table 5.Optimal geometric results

𝐃𝒓𝒄(𝐦𝐦)

𝒍𝒎

(𝐦𝐦)

𝜹(𝐦𝐦)

𝒉𝒔𝒘(𝐦𝐦)

𝒃𝒕𝒔(𝐦𝐦)

𝒉𝒔𝒔(𝐦𝐦)

𝒌𝒐𝒑𝒆𝒏

analytica

l 215 3.50 1.25 3.20 31.95 25.00

0.7280

3

f1 165.5

4 5.20 1.41 1.75 27.28 44.90

0.8609

7

f2 171.4

1 5.35 1.42 2.47 39.13 30.16

0.91731

f3 155.1

8 3.58 1.16 1.86 36.51 42.13

0.8368

8

F1 195.5

5 5.40 1.47 2.48 31.86 26.04

0.8915

1

F2 150.4

9 3.62 1.41 1.97 25.99 43.89

0.97280

F3 194.2

8 5.34 1.22 2.72 33.62 26.45

0.60763

The permanent magnet synchronous motor initially increased from

analytical design to optimization, the motor efficiency increased

from 92.33% to 94.9% while the total weight of the motor dropped

from 62.83kg to 62.46kg and the total weight of the magnets

dropped from 1.51kg to 1.1kg. While the improvement in motor

efficiency is realized in the efficiency objective function (f1), the

improvement in the total weight of the motor occurs in the

efficiency-weight objective function (F1) and the improvement in

the total weight of the magnets is in the efficiency-magnet

objective function (F2). The most striking feature is the rate of

change in the total weight of the magnets. If the results are

generally evaluated in terms of constraint functions, the magnetic

flux density does not increase so much as to cause saturation in any

region on the motor sheet.

Table 6.Optimal objective and boundary geometric results, iteration numbers and times

𝜼 (%) 𝑾𝒕𝒐𝒕(𝒌𝒈) 𝑾𝑷𝑴𝒔(𝒌𝒈) 𝑩𝒓𝒚(𝑻) 𝑩𝒔𝒚(𝑻) 𝑩𝒔𝒕(𝑻) Iter. number Iter. time (s)

analytical 92.33 62.83 1.51 0.27 1.56 1.53 - -

f1 94.90 63.58 1.76 0.34 1.09 1.54 73 16.119

f2 91.87 62.56 1.87 0.35 0.69 1.18 68 14.277

f3 92.55 63.32 1.12 0.36 0.63 1.08 18 14.467

F1 93.61 62.46 2.15 0.31 1.03 1.53 20 20.229

F2 93.54 63.29 1.10 0.32 0.53 1.29 69 15.718

F3 93.81 62.64 2.11 0.33 1.09 1.57 66 15.759

0 20 40 60 80 100

94

94.2

94.4

94.6

94.8

95

95.2

X: 73

Y: 94.9

eff

icie

ncy (

%)

iteration number

motor efficiency

0 20 40 60 80 10062.55

62.6

62.65

0 20 40 60 80 10062.55

62.6

62.65

X: 68

Y: 62.56

moto

r w

eig

ht

(kg)

0 20 40 60 80 1001.1

1.15

1.2

X: 19

Y: 1.125

magnet

weig

ht

(kg)

motor weight magnet weight

0 10 20 30 40 50 60 70 80 90 1000.93

0.94

0.95

0.96

0.97

0.98

0.99

1

X: 69

Y: 0.9403

Mem

bers

hip

valu

es

iteration number

X: 66

Y: 0.9569

X: 20

Y: 0.9903

motor efficiency & motor weight

motor efficiency & magnet weight

motor efficiency & motor weight & magnet weight

0,0

0,5

1,0

1,5

2,0

2,5

3,0

3,5

analytical f1 f2 f3 F1 F2 F3

magnet weight (kg) motor weight (kg)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 275–281 | 280

This is good news. A clear observation is that the thickness of the

rotor yoke increases all the optimization results. It is seen from the

values in Table 5, Table 6 and Fig. 7 that the most effective

geometric parameter are permanent magnets. Results of the

preliminary design and the F3 multiobjective optimization have

been tested by means of a finite element program.

According to the finite element analyses, the input power of the

motor is 2502.8W, the output power is 2251.6W, the efficiency is

89.96% and the torque is 86Nm for preliminary design. Also for

the F3 fuzzy multiobjective design optimization, the input power

of the motor is 2546.6W, the output power is 2366.9W, the

efficiency is 92.94% and the moment is 90.4Nm. Efficiency error

values are 2.63% and 0.94% for preliminary design and for the F3

fuzzy multiobjective design optimization respectively. In this case,

the fuzzy multiobjective design optimization is confirmed with a

low error and the design aims have been realized. In addition, the

magnetic flux densities obtained by a finite element program are

𝐵𝑠𝑡 = 1.14T, 𝐵𝑠𝑦 = 1.42T and 𝐵ry = 0.85T for the analytical

design and 𝐵𝑠𝑡 = 1.57T, 𝐵𝑠𝑦 = 1.28T and 𝐵ry = 1.14T for the F3

multiobjective design optimization. In this case, it can be said that

the magnetic flux densities obtained as a result of the optimization

are below the limit values and the multiobjective optimization

modelling is useful.

Motor performance is evaluated in terms of operating

performance; torque ripples of the PMSM are about 7.49% for

preliminary design and 7.16% for the F3 fuzzy multiobjective

design optimization on average, which is due to the magnet angle

and stator winding. The cogging torque is effectively 1.1Nm for

preliminary design which corresponds to 1.28% of the nominal

torque and 1.3Nm for the F3 fuzzy multiobjective design

optimization which corresponds to 1.44% of the nominal torque.

The choice of the slot number and the number of poles is influential

in the formation of this value. As a result, PMSM design

optimization with a very complex and non-linear structure has been

tried to be solved with a different approach such as fuzzy-genetic.

Taking into account the obtained optimization values, the finite

element results and the iteration times, the fuzzy-genetic approach

is quite useful for PMSM design and the designed PMSM provides

the desired performance with great precision.

5. Conclusion

In this study, design optimization of the permanent magnet

synchronous motor, which is frequently used in low speed high

torque applications, was realized by using fuzzy decision making

and genetic algorithm. Motor efficiency, motor weight and weight

of magnets were selected as the single objective functions and to

form multiobjective the single objective functions are combined

with the fuzzy decision process. Here the goal was to achieve any

single objective while other objectives were to achieve the desired

values.

Convergence curves of algorithms show performance. It can be

argued that ability of the genetic algorithm to fall locally has been

overcome by the fuzzy decision making process, which seems to

be promising. Sensitive values in the obtained design parameters

also demonstrate the ability of the algorithms to investigate. The

most successful aspect of the multiobjective algorithm is to provide

a highly flexible approach to the goals. It can be seen that both

results, fuzzy multiobjective optimization and finite element

program are very close to each other and the error margin can be

deducted from the evaluation.

References

[1] J.R. Hendershot, T.H.E. Miller (1994). Design of brushless

permanent magnet motors. Oxford University Press Inc.,

New York.

[2] D.C. Hanselman (2006). Brushless permanent magnet motor

design. Magna Physics Publishing, Ohio.

[3] S. D. Sudhoff (2014). Power magnetic devices: a multi-

objective design approach. John Wiley&Sons Inc., New

Jersey.

[4] B. N. Cassimere, S. Sudhoff (2009). Population-based

design of surface-mounted permanent-magnet synchronous

machines. IEEE Transactions on Energy Conversion, Vol.

24, No. 2, pp 338-346, doi:10.1109/TEC.2009.2016150.

[5] Y. Duan, R. G. Harley, Y. G. Habetler (2009). Comparison

of particle swarm optimization and genetic algorithm in the

design of permanent magnet motors. IEEE 6th International

Power Electronics and Motion Control Conference, pp 822-

825, doi:10.1109/IPEMC.2009.5157497.

[6] G. Zhang, M. Dou, S. Wang (2009). Hybrid genetic

algorithm with particle swarm optimization technique.

International Conference on Computational Intelligence and

Security, Vol. 1, pp 103-106, doi:10.1109/CIS.2009.236.

[7] W. H. Ho, J. T. Tsai, J. H. Chou, J. B. Yue (2016). Intelligent

hybrid taguchi-genetic algorithm for multi-criteria

optimization of shaft alignment in marine vessel. IEEE

Access, Vol. 4, pp 2304-2313,

doi:10.1109/ACCESS.2016.2569537.

[8] K. Pytel (2016). Hybrid fuzzy-genetic algorithm applied to

clustering problem. Proceedings of the Federated

Conference on Computer Science and Information Systems,

pp. 137-140, INSPEC Accession Number:16428579.

[9] A. Wang, Y. Wen, W. L. Soong, H. Li (2016). Application

of a hybrid genetic algorithm for optimal design of interior

permanent magnet synchronous machines. IEEE Conference

on Electromagnetic Field Computation, pp 1-1,

doi:10.1109/CEFC.2016.7816299.

[10] J. T. Park, C. G. Lee, M. K. Kim, H. K. Jung (1997).

Application of fuzzy decision to optimization of induction

motor design. IEEE Transactions On Magnetics, Vol. 33,

No. 2, pp 1939-1942, doi:10.1109/20.582672.

[11] E. Koskimäki, J. Göös (1997). Electric machine

dimensioning by global optimization. First International

Conference on Knowledge-Based Intelligent Electronic

Systems, doi:10.1109/KES.1997.616930.

[12] B. Mirzaeian, M. Moallem, V. Tahani, C. Lucas (2002).

Multiobjective optimization method based on a genetic

algorithm for switched reluctance motor design. IEEE

Transactions On Magnetics, Vol. 38, No. 3, pp 1524-1527,

doi:10.1109/20.999126.

[13] S. Owatchaiphong, N. H. Fuengwarodsakul (2009). Multi-

objective based optimization for switched reluctance

machines using fuzzy and genetic algorithms. International

Conference on Power Electronics and Drive Systems,

doi:10.1109/PEDS.2009.5385926.

[14] C. Choi, D. Lee, K. Park (2000). Fuzzy design of a switched

reluctance motor based on the torque profile optimization.

IEEE Transactions On Magnetics, Vol. 36, No. 5, pp 3548-

3550, doi:10.1109/20.908894.

[15] J. Pyrhonen, T. Jokinen, V. Hrabovcová (2008). Design of

rotating electrical machines. John Wiley & Sons Ltd.

[16] F. Libert (2004). Design, optimization and comparison of

permanent magnet motors for a low-speed direct-driven

mixer. Technical Licentiate, School of Computer Science,

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 275–281 | 281

Electrical Engineering and Engineering Physics, KTH,

Sweden.

[17] F. Libert, J. Soulard (2004). Design study of different direct-

driven permanent-magnet motors for a low speed

application. Proc of the Nordic Workshop on Power and

Indus Electro (NORPIE).

[18] M. Mutluer, O. Bilgin (2016). An intelligent design

optimization of a permanent magnet synchronous motor by

artificial bee colony algorithm. Turkish Journal of Electrical

Engineering & Computer Sciences, Vol. 24, pp 1826-1837,

doi:10.3906/elk-1311-150.

[19] M. Mutluer, O. Bilgin (2012) Comparison of stochastic

optimization methods for design optimization of permanent

magnet synchronous motor. Neural Computing and

Applications, Vol. 21, No 8, pp 2049-2056,

doi:10.1007/s00521-011-0627-1.

[20] M. Çunkaş (2008). Intelligent design of induction motors by

multiobjective fuzzy-genetic algorithm, Journal of

Intelligent Manufacturing, 21 (4) 393-402.

[21] A. Trebi-Ollennu, B. A. White (1997). Multiobjective fuzzy-

genetic algorithm optimization approach to nonlinear control

system design. IEE Proceedings of Control Theory and

Applications, 144 2.

[22] Y. Minghua, X. Changwen (1994). Multiobjective fuzzy

optimization of structures based on generalized fuzzy

decision making. Computers and Structures, 53(2), 411–417,

doi:10.1016/0045-7949(94)90213-5.

[23] Z. Michalewicz (1996). Genetic algorithms + data structures

= evolution programs - third, revised and extended edition.

Springer-Verlag, Berlin, Heidelberg.

[24] X. S. Yang (2010). Engineering optimization – an

introduction with metaheuristic applications. John Wiley &

Sons, Inc., Hoboken, New Jersey.

[25] S. S. Rao (2009). Engineering optimization theory and

practice – fourth edition. John Wiley & Sons, Inc., Hoboken,

New Jersey.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

This journal is © Advanced Technology & Science IJISAE, 2018, 6(4), 282–288 | 282

Surface Roughness Estimation for Turning Operation Based on

Different Regression Models Using Vibration Signals

Suleyman NESELI*1, Gokhan YALCIN2, Suleyman YALDIZ1

Accepted : 24/09/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: On machined parts, major indication of surface quality is surface roughness and also surface quality is one of the most specified

customer requirements. In the turning process, the importance of machining parameter choice is enhancing, as it controls the required

surface quality. To obtain the better surface quality, the most essential control parameters are tool overhang and tool geometry in turning

operations. The goal of this study was to develop an empirical multiple regression models for prediction of surface roughness (Ra) from

the input variables in finishing turning of 42CrMo4 steel. The main input parameters of this model are tool overhang and tool geometry

such as tool nose radius, approaching angle, and rake angle in negative direction. Regression analysis with linear, quadratic and

exponential data transformation is applied so as to find the best suitable model. The best results according to comparison of models

considering determination coefficient (R2) are achieved with quadratic regression model. In addition, tool nose radius was determined as

the most effective parameter on turning by variance analysis (ANOVA). Cutting experiments and statistical analysis demonstrate that the

model developed in this work produces smaller errors than those from some of the existing models and have a satisfactory goodness in all

three models construction and verification.

Keywords: Cutting tool geometry, Regression analysis, Surface roughness, Tool-holder overhang, Turning

1. INTRODUCTION

In many manufacturing applications, especially in the aerospace

industry, the need for long tool-holder is unavoidable. Such tools

are frequently required for the production of parts with deep holes.

The determination of optimal tool-holder overhang and cutting tool

geometry for specified surface roughness and accuracy of product

are the key factors in the selection of machining process. To

provide the quality of the process, machining processes are

producing methods, usually in relatively short periods and at low

cost. In recent years, considerable progress is made so as to

investigate the effect of tool overhang and cutting tool geometry

parameters on the resultant surface quality during single point

diamond turning [1, 2].

In recent years, due to the need to improve the quality of parts,

there has been a push toward decreasing the cutting tool holder

deflection in turning. These deflections derive from the machine

tool system and the machining process. The errors of the

machining process generated in turning originate from a number

of sources, however the tool system because of cutting force is one

of the major problems for precision machining [3]. Because the

tool holder is subject to bending depend on effect point of the

tangential cutting force (F), cutting tool displaced (). This

situation has negative effects on the surface quality as shown in

Fig.1.

Based on applications and theoretical approaches, it is known that

cutting tools must be clamped as short as possible so as to achieve

the desired surface quality of the work piece. For the internal

turning method in particular, the cutting tool should be attached

with the proper length, not with the shortest distance. This situation

may also be the case for external turning processes, depending on

the work piece geometry [4].

L

F

Tool

workpiece

Holder

Roughness

Fig. 1. Deflection of cutting tool and tool-holder, δ due to the tangential

force, F.

Kumar et al. [5] displayed the problems because of tool overhang

during boring operation. They performed particle damping based

on experiments to decrease the vibrations because of the tool

overhang and used different metal particles to fill the boring bar

for the purpose of damping. The effects of particle based damping

on surface roughness are explored. Kassab and Khoshnaw [6] have

determined the surface roughness of the work piece by selecting

tool overhangs at 25, 30, 35, and 40 mm at different cutting speeds,

different depth of cuts, and different feed rates. They studied this

during the turning of workpieces made of carbon steel materials to

determine the relevance between the tool vibration and the surface

roughness. The results showed that the feed rate had the greatest

effect on surface roughness and that both surface roughness and

vibration increased in parallel with the tool overhang. As the tool

overhang shows an increase, so does the tendency of the system to

vibrate. In all experiments, the surface roughness increased as the

tool overhang increased. Batey and Hamidzadeh [7] fulfilled

machining experiments by varying the tool feed rate, spindle speed

and tool length to measure the vibration signature for various

combinations. They transformed these signatures into the

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Mechanical Engg, Selcuk University, Konya-42002, TURKEY 2 Mechanical Drawing, Konya Teknik University, Konya-42002, TURKEY

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 282–288 | 283

frequency domain. A natural frequency of lathe components is also

calculated. Based on this study, they developed a method so as to

determine the source of vibration and for its control. According to

their results, an increase in tool length should result in higher

flexibility and lower natural frequencies. Kiyak et al. [4] studied

the effects of cutting tool overhang on the surface quality and the

cutting tool wear in external turning processes. They developed

two analytical models to find the deflection of tool at different

overhangs. The effects of tool overhang on surface roughness and

tool wear were found to be significant. Sardar et al. [8] developed

a finite element model to understand the dynamics of tool shank

overhang. They performed finite element analyses in order to

calculate the frequency at different tool overhang positions,

followed by experimental measurement of frequencies at these tool

overhang positions. It was concluded that higher tool overhang

causes more vibration in the machining system. Amin et al. [9]

have performed studies aiming to determine chip form instability

in the turning process experimentally. They used a CNMG120408

type insert and three workpiece materials (austenitic stainless steel

and AISI 1020 and AISI 1040 carbon steels). To determine the

impact of tool overhang, overhang lengths of 40, 50, and 60 mm

were selected and no significant chip form instability was observed

when the tool overhang length was 40 mm. This situation has been

referred to as the rigidity of the tool. Haddadi et al. [10] examined

the impact of worn tools and brand new tools on vibration

frequency in the turning process experimentally. They examined

the effects of rake angle and tool overhang under orthogonal and

oblique cutting conditions. In experiments, they used a workpiece

made of low carbon steel and high-speed steel tools. The tool

overhangs were 20, 30, and 50 mm. In the experimental studies

that they conducted using different depth of cuts, the effect of the

tool overhang was found to be clearly dependent on the selected

experimental parameters. Abouelatta and Mádl [11] established a

relationship between the tool life, surface roughness and the tool

vibrations. Tool overhang was selected as a variable parameter for

experiments along with the cutting speed, tool feed rate, depth of

cut, tool nose radius, approach angle, length and diameters of the

work piece. They measured the acceleration in both radial and feed

directions and the analysis of vibration was carried out by means

of Fast Fourier Transforms. Based on their experimental data,

suitable regression models were developed. Kassab and Khoshnaw

[6] have determined the surface roughness of the workpiece by

selecting tool overhangs at 25, 30, 35, and 40 mm at different

cutting speeds, different depth of cuts, and different feed rates.

They studied this during the turning of workpieces made of carbon

steel materials to determine the relevance between the tool

vibration and the surface roughness. The results showed that the

feed rate had the greatest effect on surface roughness and that both

surface roughness and vibration increased in parallel with the tool

overhang. As the tool overhang enhances, so does the tendency of

the system to vibrate. In all tests, the surface roughness increased

as the tool overhang increased.

Tool overhang effects on the quality of surface as a machining

parameter, especially during the turning process, has not been

reviewed in detail. However, few researchers have reported its

importance as one of the cutting tool geometry.

In this study, we investigate the effects of cutting tool overhang

and tool geometry on the quality of surface in external turning

processes. Three different regression models namely Linear,

quadratic and exponential have been used for evaluating their

ability to offer reasonable surface roughness prediction model and

their results are compared.

2. EXPERIMENTAL DESIGN and SETUP

2.1. Experimental setup

In the present investigation workpieces of 42CrMo4 steel, 40 mm

in diameter and 0.250 m in length, were machined using a 2.2 kW

Harrison M300 lathe. Under dry unlubricated cases, the

workpieces were machined at cutting speeds of 150 m.min-1, a feed

rate of 0.15 mm.rev-1, a depth of cut of 1.5 mm and with tool

overhangs of 30, 40 and 50 mm. The cutting tools used here are

proper for the machining of low carbon steel with ISO P25 quality,

which is equal to LC215K code. The inserts were produced by

Böhler Inc., with the ISO designation of CNMG 120404-BF,

CNMG 120408-BF, CNMG 120412-BF (80° Rhombic inserts).

The inserts were mounted rigidly on three different right hand style

tool holders designated by ISO as PCLNR/L 2020 K12 AA9,

PCLNR/L 2020 K12 AA6, and PCLNR/L 2020 K12 AA3 thus

giving back rake angle of -9°, -6° and -3°, respectively. In all

instances, the side rake angle is 6°. The experimental set up is

shown schematically in Fig. 2. After the experiments, Mahr

Perthometer M1 was used to measure the surface roughness and

the mean roughness values obtained from three different points of

machined surface were calculated.

Negative back

rake angle

Approach

angle

Nose radius

Tool overhang

limits

COMPUTER

SURFACE

ROUGHNESS

PREDICTION

PROCESS

PARAMETERS

SURFACE

RUGHNESS

MEASUREMENT

OPERATOR

STATISTICAL

SOFTWARE

Fig. 2. Schematic diagram for the experimental setup

2.2. Design of experiment

In the past, various methods were used to quantify the impact of

machining parameters on part finish quality. Although the

processes that previous researchers have utilized are similar in

nature, they all vary slightly in their execution. All of the relevant

literature includes some kind of design of experiments which allow

for a systematic approach to quantify the effects of a finite number

of parameters [12]. Most famous statistical tools contain: Taguchi

method, variance analysis, regression modelling and response

surface methodology, etc. Many researchers [13-15] used these

statistical tools to optimize parameters for different machining

processes. In the current study, for machining experiments,

maximum possible tool overhang length is divided into three

different tool overhang positions as per machine tool-holder

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 282–288 | 284

logistics. As input parameters, the total four parameters, viz: tool

overhang, tool nose radius, approach angle and rake angle are

selected accordingly, while the surface roughness is regarded as

the broad output parameter. Three different levels for each input

parameter (tool overhang, tool nose radius, approach angle and

rake angle) are selected for the machining experiments. The

parameters and their levels are summarized in Table 1.

Table 1. Process input parameters and their levels

Symbol Factor Unit Level 1 Level 2 Level 3

r Nose Radius mm 0.4 0.8 1.2

κ Approach angle Degree (°) 60 75 90

γ Rake Angle Degree (°) 9 6 3

L Tool overhang mm 30 40 50

With three tool overhang positions and three levels each of tool

nose radius, approach angle and rake angle, 81 machining cuts in

total are carried out. During each cut, the chip extraction and the

environmental conditions are kept similar. The surface roughness

was measured with a Mahr Perthometer M1, by using a trace length

of 2.5 mm, a cut-off length of 0.8 mm. For the metrology of each

machined surface, three linear scans (at 120° to each other) are

taken and their average is considered as the representative surface

roughness. Special attention is given to tool edge condition and

sufficient, special attention is given.

Variation of surface roughness-tool overhang is clear from Fig. 3.

The optimum tool overhang for machining is shown in this figure.

However, it is not giving any information about optimum tool

geometry parameters in which the tool overhang value is required

to be fixed high and very low. In such cases, the other tool

geometry parameters need to be optimized in terms of tool

overhang. This study is intended to optimize the tool geometry

parameters with respect to a fixed tool overhang position

depending on surface roughness. Hence, instead of optimizing

parameters for each tool overhang, it is decided to perform line

wise optimization. The flow chart of experiment design is shown

in Fig. 2.

Fig. 3. Effect of tool overhang on surface roughness (Ra) at tool nose

Radius 0.4 mm, approach angle 60° and rake angle -3°.

The descriptive statistics of input parameters and output

parameters (surface roughness) using in experiments and

calculated roughness values with regression analysis can be seen

in Table 2. In a data set, a smaller standard deviation indicates less

variability in ratings. Ratings with the smallest standard deviation

from the mean identified the most reliable of the three sensory

method treatments among linear quadratic, and exponential.

Table 2. Descriptive statistics of measurements

Radius App. Angle Rake Angle Overhang Ra Linear Quadratic Exponential

Mean 0.8 75 -6 40 6.4828 6.4828 6.4828 6.4686

SE Mean 0.0365 1.3693 0.2739 0.9129 0.2712 0.2577 0.2655 0.2820

Median 0.8 75 -6 40 6.563 6.4828 6.4338 6.0276

Std. Dev. 0.3286 12.3238 2.4648 8.2158 2.4410 2.3193 2.3891 2.5380

Variance 0.108 151.875 6.075 67.5 5.9586 5.3793 5.7078 6.4417

Range 0.8 30 6 20 10.315 10.6893 9.5718 11.6874

Max. 0.4 60 -9 30 2.025 1.1382 1.9967 2.2782

Min. 1.2 90 -3 50 12.34 11.8274 11.5685 13.9656

Sum 64.8 6075 -486 3240 525.107 525.107 525.107 523.9528

2.3. Methodology

To study the behaviours involved in the process, modelling is

defined as a scientific way. Regression method is one of the most

widely used statistical techniques. Multiple regression analysis is

a multivariate statistical technique used to examine the relationship

between a single dependent variable and a set of independent

variables. The objective of the multiple regression analysis is to

use independent variables whose values are known to predict the

single dependent variable. On the basis of the obtained analysis

results of the effects of process parameters on the dependent

variable and the interaction among the independent variables,

regression analysis was next carried out to estimate Ra. In turning,

there are many factors which affect the surface roughness like tool

variables, workpiece variables, and cutting conditions. Tool

variables include tool material, nose radius, rake angle, cutting

edge geometry, tool vibration, tool point angle, etc., while

workpiece variables comprise material, hardness, and other

mechanical properties. Furthermore, cutting conditions include

cutting speed, feed rate, and depth of cut. Since the turning process

contains many parameters, it is complex and hard to select the

appropriate cutting conditions and tool geometry to achieve the

required surface quality [16]. Therefore, some scientific

approaches are needed to represent the process. It seems clearly

that the proper model selection for the surface roughness is

essential for the machining of materials. Modelling of machining

processes is important to provide the basis mathematical model for

the formulation of the objective function. A model developed for

machining process is the relationship between two variables which

are decision variable and response variable in terms of

mathematical equations. Therefore, the minimization of the Ra

must be formulated in the standard mathematical model [17].

In order to accurately model the surface roughness in turning, one

needs to first understand why current models fail.

0

1

2

3

4

30 40 50

Surf

ace

Ro

ughnes

s

(µm

)

Cutting Tool Overhang (mm)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 282–288 | 285

Fig. 4. Roughness profile characteristics.

The surface roughness average Ra is generally described on the

basis of the ISO 4287 norm, which is the arithmetical mean of the

deviations of the roughness profile from the central line (lm) along

the measurement (see Fig. 4). Average roughness (Ra) of the

surface is calculated from the following equation [18].

0

1( , )aR Z x y dxdy

S (1)

where Z(X, Y) is the elevation for a given point, and S0 is the

projected area of the given area. A basic theoretical model for

surface roughness, that is, the difference between the actual surface

area and the geometrical surface area, was also calculated from the

following model: 2

32 e

fRa

r (2)

where f is feed rate and re is the tool nose radius. According to this

model, one needs only a decrease the feed rate or an increase the

tool nose radius to develop desired surface roughness. However,

there are several problems with this model. Firstly, it does not take

into account any imperfections in the process, such as tool

vibration or chip adhesion. Secondly, there are practical limitations

to this model, since certain tools (such as CBN) require specific

geometries to improve tool life [12, 19]. Many researchers have

studied the impact of cutting factors on surface roughness using

regression analysis. However, the effects of insert edge geometry

are not included in those models. There are three main effects

leading to the degradation of surface roughness: tool nose Radius,

approach angle and rake angle.

In planning and conducting the experiment, DoE 3k full factorial

were used. Selected factors of experiment have changed in 3 levels

of value. This method allows investigating the wider interval of

parameters and the predicted mathematical model is more reliable

[20]. The next step was to estimate the model.

3. RESULTS and DISCUSSION

3.1. Development of Regression Model

In this section, linear, quadric and exponential regression equations

are introduced for Ra values and model factor effects are provided.

Optimum regression model was detected according to coefficient

of determination (R2), is a number that indicates how well data fit

a statistical model. Ra is dependent variables while r, κ, γ and L are

independent variables in regression models. Linear, quadratic and

exponential regression modelling and analysis of variance

(ANOVA) for Ra values are developed using Statistica software.

The correlation coefficients (R) value of modelled independent

variable Ra is given in Tables 3, respectively.

In the next step, Ra estimation procedures were performed by

creating mathematical (linear, quadratic and exponential)

equations. When Table 3 is examined, it can be seen that the

highest correlation of R=0.9787 was obtained for Ra. Moreover,

Ra was estimated through the second degree interaction regression

equation which includes all input parameters and the determination

coefficient of R2=0.9579 was statistically significant and consistent

(Fig. 5b).

By transferring the coefficients value of independent variables

(Tables 1) into Table 4, the regression model and the determination

coefficient (R2) for Ra values can be written in the Table 4.

Table 3. Regression coefficients values for Ra

Independent

variable

Ra

coefficient

Standart

error Significant R

Fo

r li

nea

r

reg

ress

ion

mo

del

Constant -10.324 0.749 0.000

0.9501

r 5.185 0.266 0.000

0.082 0.007 0.000

-0.277 0.035 0.000

L 0.120 0.011 0.000 F

or

qu

adra

tic

reg

ress

ion

mod

el Constant -12.086 14.387 0.404

0.9787

r 27.166 16.070 0.096

0.074 0.203 0.717

-2.428 2.143 0.262

L 0.459 0.356 0.202

r2 -4.629 0.845 0.000

2 0.001 0.001 0.080

2 -0.007 0.015 0.636

L2 -0.001 0.001 0.565

r. -0.330 0.211 0.122

r. 3.234 2.471 0.195

. 0.039 0.028 0.174

r.L -0.515 0.392 0.194

.L -0.006 0.004 0.159

.L 0.076 0.052 0.150

r. . -0.058 0.033 0.078

r. .L 0.010 0.005 0.048

r. .L -0.099 0.061 0.108

..L -0.001 0.001 0.061

r. ..L 0.002 0.001 0.034

Fo

r ex

pon

enti

al

reg

ress

ion

mo

del

Constant -5.391 0.420 0.000

0.9498

r 0.673 0.029 0.000

0.959 0.080 0.000

0.266 0.029 0.000

L 0.768 0.064 0.000

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 282–288 | 286

Table 4. Linear, quadratic and Exponential effects of manually taken Ra

prediction equations.

Prediction equation R2(%)

Linear regression

model

10.324 5.186 0.082 0.277 0.12Ra r L

90.27

Quadratic

regression model

2 2 2 2

12.086 27.166 0.074 2.428 0.459

4.63 0.001 0.007 0.001

0.33 3.234 0.039 0.515 0.006 0.076

0.058 0.01 0.099 0.001 0.002

Ra r L

r L

r r rL L L

r r L r L L r L

95.79

Exponential

regression

model

0.673 0.959 0.266 0.7690.005Ra r L 90.21

Equations in Table 4 are used to calculate the predicted Ra values.

The line pattern data of Ra values of the experimental data vs.

predicted Ra values of regression model are compared in Fig. 5.

a) Comparison of the linear regression predictions and experimental alues

for Ra

b) Comparison of the quadratic regression predictions and experimental

values for Ra.

c) Comparison of the exponential regression predictions

and experimental values for Ra

In this study, experimental measured and obtained Ra values were

evaluated by using regression models that developed Statistica

software. When the Ra values are examined together, quadratic

equation gives closer results to experimental values than the other

regression models. This means that quadratic equation estimates

the surface roughness better when compared to other regression

models. By using the all values, the correlation coefficients result

in 0.9787, 0.9501 and 0.9498 for experiment- quadratic, linear and

exponential models, respectively. As the correlation coefficients

get closer to 1, estimation accuracy increases. In the case presented

in this study, the correlation coefficients obtained are very close to

1, which indicates a perfect match between regression models

estimation values and experimental measurement values (Fig. 6).

Fig. 6. The relationship between manual measurements and

the values of regression for Ra

3.2. ANOVA Analysis

For machining process ANOVA can be useful on the one hand to

determine the influence of given input parameters from a series of

experimental results by design of experiments and on the other

hand it can be used to interpret experimental data. The ANOVA

tables consist of sum of squares and degrees of freedom (DoF).

The mean-square is the ratio of sum of squares to degrees of

freedom and F ratio is the ratio of mean square to the mean square

of experimental error. In robust design F ratio can be used for

qualitative understanding of the relative factor effects. A large

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 282–288 | 287

value of F means that the effect of a given factor is large compared

to the error variance so, the larger value of F, the more important

the given factor influencing the process response.

Table 5 gives the result of the ANOVA with the surface roughness

Ra. This was performed for significance level of =0.05, i.e. for a

confidence level of 95%. A low P-value shows a statistical

significance for the source on the corresponding response [16]. The

last column of the tables shows the percentage of contribution

(PC%) of each parameter on the total variation. The greater the

percentage contribution, the greater the influence a parameter has

on the result.

Table 5. Analysis of variance for all regression models

DoF Seq SS AdjMS F-value PC%

Fo

r li

nea

r m

od

el Intercept 1 115.88 115.88 190.04 24.31

r 1 232.31 232.31 380.99 48.73

1 82.58 82.58 135.44 17.32

1 37.22 37.22 61.04 7.81

L 1 78.23 78.23 128.30 16.41

Error 76 46.34 0.61

Total 80 476.68

Fo

r qu

adra

tic

mo

del

Intercept 1 0.23 0.23 0.71 0.05

r 1 0.94 0.94 2.86 0.20

1 0.04 0.04 0.13 0.01

1 0.42 0.42 1.28 0.09

L 1 0.55 0.55 1.66 0.11

r2 1 9.88 9.88 30.03 2.07

2 1 1.04 1.04 3.16 0.22

2 1 0.07 0.07 0.23 0.02

L2 1 0.11 0.11 0.33 0.02

r. 1 0.81 0.81 2.45 0.17

r. 1 0.56 0.56 1.71 0.12

. 1 0.62 0.62 1.89 0.13

r.L 1 0.57 0.57 1.73 0.12

.L 1 0.67 0.67 2.04 0.14

.L 1 0.70 0.70 2.12 0.15

r. . 1 1.06 1.06 3.21 0.22

r. .L 1 1.34 1.34 4.08 0.28

r. .L 1 0.87 0.87 2.66 0.18

..L 1 1.20 1.20 3.65 0.25

r. ..L 1 1.54 1.54 4.68 0.32

Error 61 20.06 0.33

Total 80 476.68

Fo

r ex

pon

enti

al m

odel

Intercept 1 2.36 2.36 164.52 16.92

r 1 7.55 7.55 526.01 54.08

1 2.05 2.05 142.56 14.66

1 1.18 1.18 82.31 8.46

L 1 2.09 2.09 145.71 14.98

Error 76 1.09 0.01

Total 80 13.96

From the analysis in Table 5, the most significance factor for F-

value is the linear and quadratic effect of tool nose radius with

380.99, 30.03, 526.01 F ratio for linear, quadratic and exponential

regression models, respectively. Also there isn’t a strong relation

between independent variables and dependent variable because of

low determination coefficient value of linear and exponential

regression models comparing with quadratic model determination

coefficient value. This is a good agreement with the previous

researcher’s works [21, 22]. Here are the cutting parameters which

mainly control the final result and especially the roughness of the

machined surface. This is the technical result that is usually

searched.

3.3. Graphical Analysis

The response surface is plotted (Fig. 7a–f) so as to search the effect

of process variables on the end of surface. The response surface

plots represent the regression equations which are used to describe

the relationship between the response and experimental levels of

each parameter. From Fig. 7a, c, e, the surface roughness Ra has

an increasing trend with the increase of radius, overhang and

approach angle. From Fig. 7b, d, f it is seen that, Ra decreases with

the decreases of depth of nose radius. It can be noted that from Fig.

7, tool nose radius has the most effect on surface roughness and its

variation is very high when compared to other parameters. The best

surface roughness was achieved at the lowest radius, approach

angle and overhang and highest rake angle combination as

expected. This is an established fact that more vibrations are

resulted from high value of tool overhang and nose radius.

Similarly, with the increasing approach angle, the cutting force

increases as well. Thus, at this time vibration is born.

Fig. 7. 3D Response surface plots for Ra vs. input factors

4. CONCLUSION

In this study, the behaviour of tool overhang and cutting tool

geometry in single turning point and its influence on the quality of

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 282–288 | 288

surface through a series of its interaction with the other machining

parameters (viz: tool nose radius, approach angle and rake angle)

are investigated. A series of machining operations are performed

in the sequential combinations of tool overhang, tool nose radius,

approach angle and rake angle. The surface quality is analysed in

terms of linear, quadratic and exponential regression analysis and

ANOVA. The regression analyses are used to predict the surface

roughness by means of suitable modelling operations. The

followings are the conclusions based on the study above:

In single point turning, the tool overhang influences the quality

of surface significantly. Extreme values of tool overhang result

in poor surface quality. From the experiments carried out on

the machining parameters, it was observed that the surface

roughness of work piece increases as the tool overhang

increases. The optimum range of tool overhang is 30 mm.

Using the same tool overhang, the surface roughness of the

work piece shows an increase as the tool nose radius and

approach angle increases. In the measurements performed after

the experiments were complete, it was seen that the cutting tool

deflection values increased together with the tool overhang.

Also, strong interaction among all input process parameters is

observed and validated. Thus, the selection of the other

machining parameters are affected by the tool overhang as well

considerably.

Rake angle parameter adversely affects surface finish.

Equations which were derived by multi-regression modelling

techniques can be used for surface roughness modelling of

turning process that can’t be computed by theoretically.

Surface roughness value can be predicted with an acceptable

accuracy with the usage of linear, quadratic and exponential

regression models. The result of comparison between these

models and coefficient of correlations (R2) for predicted

surface roughness shows that prediction performance of the

quadratic model is slightly better than the others.

In the future, real time monitoring, optimization, model

referenced adaptive control of the process can be studied while

different parameters would be added.

References

[1] Mishra, V., Khan, G.S., Chattopadhyay, K.D., Nand, K., and

Sarepaka, R.G.V., “Effects of tool overhang on selection of

machining parameters and surface finish during diamond

turning”, Measurement 55, 353-361, (2014).

[2] Kotkar, D.R., Wakchaure, V.D., “Vibration control of newly

designed Tool and Tool-Holder for internal treading of

Hydraulic Steering Gear Nut”, International Journal Of

Modern Engineering Research 4(6), 46-57, (2014).

[3] Hadi, Y., “Dynamic Deflection of Periodic Cutting Tool

Holder Based on Passive Model”, International Journal of

Mechanical & Mechatronics Engineering 11 No: 06, (2011).

[4] Kiyak, M., Kaner, B., Sahin, I., Aldemir, B., and Cakir, O.,

“The Dependence of Tool Overhang on Surface Quality and

Tool Wear in the Turning Process”, International Advanced

Manufacturing Technology 51(5), 431-438, (2010).

[5] Sathishkumar, B., Mohanasundaram, K.M. and Senthilkumar,

M., “Experimental studies on impact of particle damping on

surface roughness of machined components in boring

operation”, European Journal of Scientific Research 71(3),

327-337, (2012).

[6] Kassab, S.Y., and Khoshnaw, Y.K., “The effect of cutting tool

vibration on surface roughness of workpiece in dry turning

operation”, International Journal of Engineering Science and

Technology 25(7), 879-889, (2007).

[7] Batey, M.C., and Hamidzadeh, H.R., “Turning process using

vibration signature analysis”, Journal of Vibration and

Control 13(5), 527-536, (2007).

[8] Sardar, N., Bhaumik, A., Mandal, N.K., “Modal analysis and

experimental determination of optimum tool shank overhang

of a lathe machine”, Sensors & Transducers Journal 99(12),

53-65, (2008).

[9] Amin, A.K.M.N., Nashron, F.R. and Zubaire, W.W.D., “Role

of the frequency of secondary serrated teeth in chatter

formation during turning of carbon steel AISI 1040 and

stainless steel”, Proceeding 1st International Conference &

7th AUN/SEED-Net, Field-wise Seminar on Manufacturing

and Material Processing, 181-186, (2006).

[10] Haddadi, E., Shabghard, M.R. and Ettefagh, M.M., “Effect

of different tool edge conditions on wear detection by

vibration spectrum analysis in turning operation”, Journal of

Applied Sciences 8(21), 3879-3886, (2008).

[11] Abouelatta, O.B. and Mádl, J., “Surface roughness

prediction based on cutting parameters and tool vibrations in

turning operations”, Journal of Materials Processing

Technology 118, 269-277, (2001).

[12] Özel, T. and Karpat, Y., “Predictive modeling of surface

roughness and tool wear in hard turning using regression and

neural networks”, International Journal of Machine Tools &

Manufacture 45, 467-479, (2005).

[13] Adalarasan, R., Santhanakumar, M. and Rajmohan, M.,

“Optimization of laser cutting parameters for

Al6061/SiCp/Al2O3 composite using grey based response

surface methodology (GRSM)”, Measurement, 73, 596-606,

(2015).

[14] Debnath, S., Reddy, M.M. and Yi, Q.S., “Influence of cutting

fluid conditions and cutting parameters on surface roughness

and tool wear in turning process using Taguchi method”,

Measurement, 78, 111-119, (2016).

[15] Asiltürk, İ., Neşeli, S. and İnce, M.A., “Optimisation of

parameters affecting surface roughness of Co28Cr6Mo

medical material during CNC lathe machining by using the

Taguchi and RSM methods”, Measurement, 78, 120-128,

(2016).

[16] Singh, D. and Rao, P.W., “A Surface roughness prediction

model for hard turning process”, International Journal of

Advanced Manufacturing Technology, 32(11–12), 1115–

1124, (2007).

[17] Zain, A.M., Haron, H., Qasem, S.N. and Sharif, S.,

“Regression and ANN models for estimating minimum value

of machining performance”, Applied Mathematical

Modelling, 36, 1477-1492, (2012).

[18] Arbizu, P.I. and Pérez, C.J.L., “Surface roughness prediction

by factorial design of experiments in turning processes”,

Journal of Materials Processing Technology, 143–144, 390-

396, (2003).

[19] Thiele, J.D., Melkote, S.N., Peascoe, R.A. and Watkins,

T.R., “Effect of cutting-edge geometry and workpiece

hardness on surface residual stresses in finish hard turning of

AISI 52100 steel”, ASME Journal of Manufacturing Science

and Engineering, 122, 642-649, (2000).

[20] Montgomerty, D.C., “Design and Analysis of Experiments”,

eighth ed., John Willey&Sons Inc., New York, (2013).

[21] Asilturk, I., Celik, L., Canlı, E., and Onal, G., “Regression

Modeling of Surface Roughness in Grinding”, Advanced

Materials Research, 271(273), 34-39, (2011).

[22] Tasdemir, S., Urkmez, A., and Inal, S., “Determination of

body measurements on the Holstein cows using digital image

analysis and estimation of live weight with regression

analysis”, Computers and Electronics in Agriculture, 76,

189-197, (2011).

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 289–293 | 289

Breast Cancer Diagnosis by Different Machine Learning Methods Using

Blood Analysis Data

Muhammet Fatih Aslan*1, Yunus Celik1, Kadir Sabanci1, Akif Durdu2

Accepted : 18/11/2018 Published: 31/12/201x DOI: 10.1039/b000000x

Abstract: Today, one of the most common types of cancer is breast cancer. It is crucial to prevent the propagation of malign cells to reduce

the rate of cancer induced mortality. Cancer detection must be done as early as possible for this purpose. Machine Learning techniques are

used to diagnose or predict the success of treatment in medicine. In this study, four different machine learning algorithms were used to

early detection of breast cancer. The aim of this study is to process the results of routine blood analysis with different ML methods and to

understand how effective these methods are for detection. Methods used can be listed as Artificial Neural Network (ANN), standard

Extreme Learning Machine (ELM), Support Vector Machine (SVM) and K-Nearest Neighbor (k-NN). Dataset used were taken from UCI

library. In this dataset age, body mass index (BMI), glucose, insulin, homeostasis model assessment (HOMA), leptin, adiponectin, resistin

and chemokine monocyte chemoattractant protein 1 (MCP1) attributes were used. Parameters that have the best accuracy values were

found by using four different Machine Learning techniques. For this purpose, hyperparameter optimization method was used. In the end,

the results were compared and discussed.

Keywords: Breast cancer, Artificial Neural Network, Extreme Learning Machine, Support Vector Machine, K-Nearest Neighbors,

Hyperparameter Optimization

1. Introduction

Different cancer types have long been a major threat to human life

[1]. Among these types, breast cancer has a high mortality rate in

women. Unfortunately, this rate is increasing in developed

countries day by day [2, 3]. Moreover, breast cancer is the second

biggest cause of death all over the world [4]. According to World

Health Organization (WHO) data [5], breast cancer has been

detected in 25% of women in the United Nations [6]. 16% of all

female cancers is breast cancer [5].

Cancer is a sickness that starts in the cell and spreads into the other

part of the body [7]. That is the reason why early detection is

crucial to prevent before it spreads. Early diagnosis of breast

cancer is the most important and difficult part of breast imaging

[8]. The works for early detection of breast cancer are not new but

the current works are not capable enough for early detection so, in

addition to current works, scientist are searching for new methods

[9]. Specially Computer-Aided Detection (CAD) systems play a

crucial role in early detection [10].

Machine Learning (ML) techniques are used in the CAD system

applications. ML [11] is an Artificial Intelligence (AI) topic that

enables the machines to learn a special task by experience. In

recent years, ML methods have become widespread in predicting

and detecting applications in order to make strong decisions in

recent years. For example, ML methods can be used to determine

whether a cancer is benign or malign [9].

2. Related Works

There are many works exist for the detection of breast cancer using

ML techniques in the literature. In this part, some of these works

were shown. The performance comparison of Support Vector

Machine (SVM), K-Nearest Neighbor (k-NN), Decision Tree

(C4.5) and Naive Bayes (NB) Machine Learning (ML) techniques

were shown [12]. Wisconsin Diagnosis Breast Cancer (WDBC)

dataset [13] was used for this work. The best result was obtained

with SVM technique as 97.13%. The paper with [14] reference

number includes a work that K-Means and SVM algorithms were

used as a hybrid for the purpose of detection of a tumour. A

classification with a high accuracy rate was performed as a result

of 10 times cross-validation. In this work, the WDBC dataset [13]

was used. 97.38% accuracy was achieved. The paper [15] shows

the success of SVM and Artificial Neural Network (ANN)

techniques together. WDBC dataset was also used in this paper.

Accuracy was obtained as 97.14% with SVM and 96.71% with

ANN. According to these results, SVM gives better results than

ANN. Another paper [16] it was also shown that SVM has better

performance for the detection of breast cancer. On the other hand,

the performance of the SVM depends on the kernel function. In

this paper, the performance of different types of kernel functions

were compared. In the paper [17], k-NN algorithm was optimized

for a faster and more reliable classification. 94.1% accuracy was

obtained. The paper [18] is about the usage of different ML

techniques for breast cancer. The research is about the usage of

ANN, SVM, Decision Tree (DT) and k-NN techniques in breast

cancer diagnosis. In the paper [19] DT, Bayesian Belief Network,

and SVM techniques were compared. In the last paper [20], breast

cancer was detected using ANN classification. The work focuses

on the optimal activation function that minimizes the classification

error by using fewer blocks.

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1Department of Electrical-Electronic Engineering, Karamanoğlu

Mehmetbey University,Karaman,-70100, Turkey 2Department of Electrical-Electronic Engineering, Konya Technical

University,Konya,-42000, Turkey,

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 289–293 | 290

3. Material and Methods

In this section, dataset and ML methods used were presented.

3.1. Data Understanding

When related works were analysed, it is clear that there are several

different techniques for the detection of breast cancer and detection

problem still exist. There are several types of the dataset for the

detection of breast cancer. In this paper, Breast Cancer Coimbra

dataset [22] taken from UCI [21] ML Repository was used. This

dataset includes features that can be collected in routine blood

analysis. These features are age (years), BMI (kg/m2), Glucose

(mg/dL), Insulin (µU/mL), HOMA, Leptin (ng/mL), Adiponectin

(µg/mL), Resistin (ng/mL) and MCP1(pg/dL). According to these

input features, target data can be classified as healthy or unhealthy.

These features were measured from 64 patients with breast cancer

and 52 healthy people [22,23]. This dataset differs from others in

terms of the features it contains.

3.2. Artificial Neural Network (ANN)

The structure of the ANN is quite similar to biological neural

networks [24]. An ANN composed of three layers as an input layer,

hidden layer, and an output layer. The neurons in each layer are

connected to each other with a specific weight. These weights are

updated themselves iteratively until they are close enough to target

values. When the weights are tuned, the system can be expressed

as trained. After this phase, the testing process can be performed

[25].

3.3. Extreme Learning Machine (ELM)

ELM is a method invented by Huang and friends [26]. It is actually

having the same structure with ANN. The difference is that while

ANN has more than one hidden layer, standard ELM should have

only one hidden layer. Moreover, Unlike ANN, there are more than

1000 hidden layer neurons in a standard ELM [27]. ELM offers

advantages over other ML methods in terms of speed. Because

ELM completes training with a single iteration [28]. Weights are

assigned randomly and according to target values, β values are

calculated. Moore-Penrose generalized inverse matrix method is

used for the calculation [29].

3.4. Support Vector Machine (SVM)

SVM is one of the best advised ML methods in terms of speed and

accuracy [30]. SVM forms optimal hyperplanes in a

multidimensional plane and in this way classifies multi-class

property data [31]. The SVM contains calculations for the creation

of this plane. If the properties can be classified as linear, the plane

can be created by simple calculations. The kernel trick is used for

non-linear features. With kernel trick, features can be converted to

a higher-level and can be separated linearly [32].

3.5. K-Nearest Neighbors (k-NN)

With the help of k-NN, data in the feature space are classified

according to distance. The distances can be calculated with

different methods. In order to classify the data, the decision is made

by looking at the distance from k numbered neighbor. Data is

assigned to the nearest class [33]. Since there is no training phase,

understanding and implementation of the method are quite easy

[34].

4. Application and Results

In this study, the dataset was taken from the UCI library [21] and

blood analysis data taken from the paper [22] were used. There are

116 samples in total. Some of these data are shown in Table 1.

Target values indicate that the person is healthy or unhealthy.

Considering the input values, max and min of these values are quite

different from each other. Normalization must first be applied to

normalize the distribution and increase the success rate.

Feature Scaling method is used for normalization. The formula for

this method is shown in (1).

𝑥′ =𝑥 − 𝑥𝑚𝑖𝑛

𝑥𝑚𝑎𝑥 − 𝑥𝑚𝑖𝑛 (1)

After normalization using (1), training and test data were generated

randomly from the data. 80% percent of the whole data were used

in the test phase and 20% percent were used in the training phase.

After separation of training and test data, results were obtained for

each ML method.

An interface was created in MATLAB GUI environment for

classification with ANN (see Fig. 1). In ANN, there are a number

of hyperparameters that affect the accuracy of the system. The

important parameters can be listed as Number of Hidden Layer

Neuron, Epoch Number, Learning Rate and Momentum

Coefficient. These values must be set by trial and error to obtain

the most optimal result of ANN. For this reason, at the interface, a

certain range of these parameters can be adjusted by the user. The

graphs in the interface give Root Mean Square Error (RMSE)

values according to the changing parameter values. The results of

the RMSE values are plotted according to the changed parameter

and the parameters which give the minimum error were recorded.

After that, training and test process was managed by using the best

parameters. As a result, the average test accuracy rate is 79.4304%,

average training time 0.4282 second and average RMSE value was

obtained as 0.3954. Comparison of these values with ELM is

shown in Table 2.

Table 1. Some data used for breast cancer detection

Age BMI Glucose Insulin HOMA Leptin Adiponectin Resistin MCP.1 Class

48 23.5 70 2.707 0.4674087 8.8071 9.7024 7.99585 417.114 1

83 20.690495 92 3.115 0.7068973 8.8438 5.429285 4.06405 468.786 1

82 23.12467 91 4.498 1.0096511 17.9393 22.43204 9.27715 554.697 1

68 21.367521 77 3.226 0.6127249 9.8827 7.16956 12.766 928.22 1

45 21.303949 102 13.852 3.4851632 7.6476 21.056625 23.0341 552.444 2

45 20.829995 74 4.56 0.832352 7.7529 8.237405 28.0323 382.955 2

49 20.956607 94 12.305 2.8531193 11.2406 8.412175 23.1177 573.63 2

34 24.242424 92 21.699 4.9242264 16.7353 21.823745 12.0653 481.949 2

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 289–293 | 291

Fig. 1. Designed ANN interface and results

An interface was created in the MATLAB GUI environment for

standard ELM classification (see Fig. 2). In standard ELM, the

hyperparameter that affects the accuracy of the system is the

number of hidden layer neuron. The number of hidden layer

neurons is changed within a certain range to achieve the most

optimal result with ELM. This range can be determined by the user.

RMSE values are plotted according to the changed parameter. The

best number of hidden neuron layers was obtained as 1800 (see

Fig. 2). As a result, the average test accuracy rate is 80%, average

training time is 0.0075 second and average RMSE value was

obtained as 0.4755. Comparison of these values with ANN is

shown in Table 2.

Fig. 2. Designed ELM interface and results

Table 2 shows that the accuracy values of ANN and standard ELM

are close to each other. But ELM is much faster than ANN. When

number of training samples is too high, the use of standard ELM is

much more advantageous in terms of time.

Hyperparametric optimization method is also used for

classification with k-NN. These parameters can be thought of as

the number of neighbors and distance type for k-NN. The resulting

Hyperparameter optimization in MATLAB environment is shown

in Fig. 3. Optimum parameter values are determined according to

the graph. Euclidean distance was used as a distance type. The

number of neighbors was chosen as two. Average accuracy rate

was obtained as 77.5%. Using the best parameters, the training data

was classified at 0.15781 sec.

Fig. 3. Hyperparameter Optimization for k-NN algorithm

Hyperparameter optimization is also used for classification with

SVM. Hyperparameters of SVM can be thought as regularization

constant (box Constraint (C)) and kernel scale for SVM. The soft

margin method has been taken into account in the classification by

SVM. The resulting Hyperparameter optimization in MATLAB

environment is shown in Fig. 4. Optimum parameter values are

determined according to this graph. Optimal kernel scale value

found 0.0287. Optimum C value was obtained as 0.4869. As a

Table 2. Comparison of ANN and ELM results for 10 data

ELM ANN

Data

Number

Train Test Train Test

Acc. Rate Train Time Acc. Rate RMSE Acc. Rate Train Time Acc. Rate RMSE

1 83.8710 0.0073 78.2609 0.4802 76.3441 0.4620 78.2610 0.3618

2 83.8710 0.0136 82.6087 0.4706 70.9677 0.3797 78.2610 0.3952

3 83.8710 0.0060 82.6087 0.4631 80.6452 0.3864 73.9130 0.3858

4 83.8710 0.0079 73.9130 0.4863 75.2688 0.5253 69.5652 0.4961

5 83.8710 0.0060 78.2610 0.4817 74.1935 0.3666 82.6087 0.3820

6 83.8710 0.0063 82.6087 0.4814 83.8710 0.4406 82.6087 0.3932

7 84.9462 0.0069 78.2610 0.4634 82.7957 0.4107 89.9565 0.3527

8 84.9462 0.0072 82.6087 0.4820 80.6452 0.4761 82.6087 0.3733

9 83.8710 0.0068 82.6087 0.4701 74.1935 0.4105 73.913 0.4108

10 81.7204 0.0073 78.2610 0.4759 78.4946 0.4244 82.6087 0.4034

Average 83.8710 0.0075 80.00 0.4755 77.7419 0.4282 79.4304 0.3954

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 289–293 | 292

result, average accuracy rate was obtained as 73.5%. Using the best

parameters, the training data was classified at 0.1866 sec.

Fig. 4. Hyperparameter Optimization for SVM algorithm

5. Conclusion and Discussion

In this study, Breast Cancer Coimbra dataset [22] taken from UCI

[21] was used. This dataset is different from other datasets in terms

of feature type. This dataset includes age, BMI, glucose, insulin,

HOMA, leptin, adiponectin, resistin and MCP1 features that can

be collected in routine blood analysis. The significance of these

data in breast cancer detection was investigated by ML methods.

Analysis was performed with four different ML methods.

Interfaces for ANN and ELM have been developed. In addition,

the hyperparameter values giving the least errors for ANN, ELM,

k-NN and SVM methods are determined using Hyperparameter

optimization technique. Accuracy rates and training times were

obtained according to these values. Calculated accuracy values and

training time are shown in Table 3. The k-NN method does not

actually contain the training phase. The value in Table 3 represents

the calculation period of the training data.

Table 3. Comparison of ML algorithms

ML

Algorithms ANN ELM k-NN SVM

Acc. Rate

(%) 79.4304 80 77.5 73.5

Train

Time (sec) 0.4282 0.0075 0.15781 0.1866

When the values in Table 3 are examined, the highest accuracy rate

and the lowest training period are provided by standard ELM.

According to these results, the use of standard ELM is more

advantageous in terms of time when there is a high number of

samples. The importance of this work is pretty high because of the

usage of the different type of data. In addition, this study is also

important because four different ML methods are compared. As a

result of the study, the obtained accuracy rate cannot be regarded

as very high. However, this study investigated the utility of such

data with ML methods in breast cancer detection. In addition, this

study may support the further work in this field.

References

[1] J. Tang, R. M. Rangayyan, J. Xu, I. El Naqa, and Y. Yang,

“Computer-aided detection and diagnosis of breast cancer with

mammography: recent advances,” IEEE Transactions on

Information Technology in Biomedicine, vol. 13, no. 2, pp. 236-251,

2009.

[2] Z. Ahmad, A. Khurshid, A. Qureshi, R. Idress, N. Asghar, and N.

Kayani, “Breast carcinoma grading, estimation of tumor size,

axillary lymph node status, staging, and nottingham prognostic

index scoring on mastectomy specimens,” Indian Journal of

Pathology and Microbiology, vol. 52, no. 4, pp. 477, 2009.

[3] U. R. Acharya, E. Y.-K. Ng, J.-H. Tan, and S. V. Sree,

“Thermography based breast cancer detection using texture features

and support vector machine,” Journal of medical systems, vol. 36,

no. 3, pp. 1503-1510, 2012.

[4] K. Ganesan, U. R. Acharya, C. K. Chua, L. C. Min, K. T. Abraham,

and K.-H. Ng, “Computer-aided breast cancer detection using

mammograms: a review,” IEEE Reviews in biomedical engineering,

vol. 6, pp. 77-98, 2013.

[5] WHO. “Breast cancer: prevention and control,”

http://www.who.int/cancer/detection/breastcancer/en/index1.html.

[6] D. Bazazeh, and R. Shubair, “Comparative study of machine

learning algorithms for breast cancer detection and diagnosis.” 5th

International Conference on Electronic Devices, Systems and

Applications (ICEDSA), pp. 1-4, 2016.

[7] M. Hejmadi, Introduction to cancer biology, 2 ed.: Bookboon, 2009.

[8] I. Schreer, and J. Lüttges, “Breast cancer: early detection,”

Radiologic-Pathologic Correlations from Head to Toe, pp. 767-784:

Springer, 2005.

[9] B. K. Gayathri, and P. Raajan, “A survey of breast cancer detection

based on image segmentation techniques.” IEEE International

Conference on Computing Technologies and Intelligent Data

Engineering (ICCTIDE'16), pp. 1-5, 2016.

[10] P. Král, and L. Lenc, “LBP features for breast cancer detection.”

IEEE International Conference on Image Processing (ICIP), pp.

2643-2647, 2016.

[11] P. Louridas and C. Ebert, “Machine Learning,” IEEE Software, vol.

33, no. 5, pp. 110-115, 2016.

[12] H. Asri, H. Mousannif, H. Al Moatassime, and T. Noel, “Using

machine learning algorithms for breast cancer risk prediction and

diagnosis,” Procedia Computer Science, vol. 83, pp. 1064-1069,

2016.

[13] K. P. Bennett, and J. A. Blue, “A support vector machine approach

to decision trees.” IEEE International Joint Conference on Neural

Networks Proceedings. IEEE World Congress on Computational

Intelligence, pp. 2396-2401, 1998.

[14] B. Zheng, S. W. Yoon, and S. S. Lam, “Breast cancer diagnosis

based on feature extraction using a hybrid of K-means and support

vector machine algorithms,” Expert Systems with Applications, vol.

41, no. 4, pp. 1476-1482, 2014.

[15] R. Alyami, J. Alhajjaj, B. Alnajrani, I. Elaalami, A. Alqahtani, N.

Aldhafferi, T. O. Owolabi, and S. O. Olatunji, “Investigating the

effect of Correlation based Feature Selection on breast cancer

diagnosis using Artificial Neural Network and Support Vector

Machines.” IEEE International Conference on Informatics, Health

& Technology (ICIHT), pp. 1-7, 2017.

[16] M. Hussain, S. K. Wajid, A. Elzaart, and M. Berbar, “A comparison

of SVM kernel functions for breast cancer detection.” IEEE Eighth

International Conference Computer Graphics, Imaging and

Visualization, pp. 145-150, 2011

[17] W. Cherif, “Optimization of K-NN algorithm by clustering and

reliability coefficients: application to breast-cancer diagnosis,”

Procedia Computer Science, vol. 127, pp. 293-299, 2018.

[18] W. Yue, Z. Wang, H. Chen, A. Payne, and X. Liu, “Machine

Learning with Applications in Breast Cancer Diagnosis and

Prognosis,” Designs, vol. 2, no. 2, pp. 13, 2018.

[19] P. Suryachandra, and P. V. S. Reddy, “Comparison of machine

learning algorithms for breast cancer.” IEEE International

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 289–293 | 293

Conference on Inventive Computation Technologies (ICICT), pp. 1-

6, 2016.

[20] H. Jouni, M. Issa, A. Harb, G. Jacquemod, and Y. Leduc, “Neural

Network architecture for breast cancer detection and classification.”

IEEE International Multidisciplinary Conference on Engineering

Technology (IMCET), pp. 37-41, 2016.

[21] UCI. “Machine Learning Repository,”

https://archive.ics.uci.edu/ml/index.php.

[22] M. Patrício, J. Pereira, J. Crisóstomo, P. Matafome, M. Gomes, R.

Seiça, and F. Caramelo, “Using Resistin, glucose, age and BMI to

predict the presence of breast cancer,” BMC cancer, vol. 18, no. 1,

pp. 29, 2018.

[23] J. Crisóstomo, P. Matafome, D. Santos-Silva, A. L. Gomes, M.

Gomes, M. Patrício, L. Letra, A. B. Sarmento-Ribeiro, L. Santos and

R. Seiça, “Hyperresistinemia and metabolic dysregulation: a risky

crosstalk in obese breast cancer,” International Journal of Basic and

Clinical Endocrinology, vol. 53, no. 2, pp. 433-442, 2016.

[24] P. Kshirsagar, and N. Rathod, “Artificial neural network,”

International Journal of Computer Applications, 2012.

[25] N. Gupta, “Artificial neural network,” Network and Complex

Systems, vol. 3, no. 1, pp. 24-28, 2013.

[26] G.-B. Huang, Q.-Y. Zhu, and C.-K. Siew, “Extreme learning

machine: a new learning scheme of feedforward neural networks.”

IEEE International Joint Conference on Neural Networks, pp. 985-

990, 2004.

[27] Y. Yang, and Q. M. J. Wu, “Extreme Learning Machine With

Subnetwork Hidden Nodes for Regression and Classification,” IEEE

Transactions on Cybernetics, vol. 46, no. 12, pp. 2885-2898, 2016.

[28] S. Ding, H. Zhao, Y. Zhang, X. Xu, and R. Nie, “Extreme learning

machine: algorithm, theory and applications,” Artificial Intelligence

Review, vol. 44, no. 1, pp. 103-115, 2015.

[29] N.-Y. Liang, G.-B. Huang, P. Saratchandran, and N. Sundararajan,

“A fast and accurate online sequential learning algorithm for

feedforward networks,” IEEE Transactions on neural networks, vol.

17, no. 6, pp. 1411-1423, 2006.

[30] V. N. Mandhala, V. Sujatha, and B. R. Devi, “Scene classification

using support vector machines.” IEEE International Conference on

Advanced Communications, Control and Computing Technologies,

pp. 1807-1810, 2014.

[31] M. E. Mavroforakis, and S. Theodoridis, "Support Vector Machine

(SVM) classification through geometry.", IEEE 13th European

Signal Processing Conference, pp. 1-4, 2005.

[32] W.-C. Lai, P.-H. Huang, Y.-J. Lee, and A. Chiang, “A distributed

ensemble scheme for nonlinear support vector machine,” IEEE

Tenth International Conference on Intelligent Sensors, Sensor

Networks and Information Processing (ISSNIP), pp. 1-6, 2015.

[33] N. Suguna, and K. Thanushkodi, “An improved k-nearest neighbor

classification using genetic algorithm,” International Journal of

Computer Science Issues, vol. 7, no. 2, pp. 18-21, 2010.

[34] J. Kim¹, B.-S. Kim, and S. Savarese, “Comparing image

classification methods: K-nearest-neighbor and support-vector-

machines,” Proceedings of the 6th WSEAS international conference

on Computer Engineering and Applications, and Proceedings of the

2012 American conference on Applied Mathematics, vol. 1001, pp.

48109-2122, 2012.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 294–298 | 294

FACE VERIFICATION SYSTEM IN MOBILE DEVICES BY USING

COGNITIVE SERVICES

Adem Alpaslan ALTUN*1, Cagatay KOLUS2

Accepted : 19/10/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: Biometric systems enable people to distinguish between physical and behavioral characteristics. Face recognition systems, a

type of biometric systems, use peoples’ facial features to recognize them. The aim of this study is to perform face recognition and

verification system that can run on mobile devices. The developed application is based on comparing the faces in two photographs. The

user uploads two photos to the system, the system identifies the faces in these photos and performs authentication between the two faces.

As a result, the system gives the output that the two faces in the photo belong to the same or different persons. It provides a security measure

thanks to the face identification and verification feature included in this application. This application can be integrated into various

applications and used in systems such as user login.

Keywords: Cognitive services, Face identification, Face verification, Image Processing, Mobile application

1. Introduction

Facial recognition technology is used in many different areas today

[1-3]. The use of the system as an input method, detection of

criminals can be given as an example of some of these areas of use

[4, 5]. According to the researchers, the face recognition method is

seen as a more natural and effective biometric method than the

identification methods such as iris, fingerprint [6].

Facial recognition technology provides a variety of output for the

inputs that are trained as a result of the system being trained. The

success of the system depends on the most effective way of

learning, and on the success of various processes applied to inputs

[7]. Face recognition is based on calculating the ratio of the face to

be recognized to the faces in the database. This is accomplished by

finding that the queried photo matches more closely with the

photos in the database [8].

2. Materials and Method

Mobile applications are software programs developed for mobile

devices such as smartphones and tablets [9, 10]. In this study,

Android Studio application development platform and Microsoft

Cognitive Services were used. Android Studio is a mobile

application development environment that provides features such

as code editing, debugging, and performance tracking, enabling

applications to run on any Android device [11].

Microsoft Cognitive Services are collection of machine learning

algorithm which helps in solving various problems in the field of

artificial intelligence, like language processing, machine learning

search, computer vision etc. Basically Cognitive Services are

collection of APIs, SDKs and services designed for developers.

These services can make the applications more intelligent and

more interactive. The aim of these services is to supply interesting

and well - off computing experience. The available APIs of

Microsoft Cognitive Services are Language API, Vision API,

Speech API, Knowledge API , etc . Each API performs different

functions such as language API identifies and discovers the

requirements of the user , vision API examines the images an d

videos for useful information , speech API helps in identifying the

speaker and knowledge API captures research from scientific

account. By using these APIs developers can add the intelligent

features like understanding face detection, speech detection, vision

detection and recognition, emotion detection and video detection.

The characteristics which distinguish Microsoft Cognitive

Services from other services are multiple face tracking in less time,

more accuracy in face recognition, presence of emotions with their

types and percentages , better APIs. Microsoft Cognitive Service

APIs are used in various field s like enhancing the security,

expressing farcical moments, engaging customers via chat, etc.

[12]

Microsoft Cognitive Services helps developers create intelligent

applications. It provides many features for developers such as

speech and image recognition, face and emotion detection [13].

This API performs functions such as detecting face attributes and

recognizing the face. It basically helps in detecting, analyzing and

organizing in given images. This API detects human face with high

precision. It allows us to tag faces in any given photo.

Alternatively, face detection allows extraction of face attributes

like gender, age, facial hair, glasses, head pose etc. This API

provides four face recognition func tions: face verification, finding

similar faces, face grouping, and identification. This API helps in

adding super cool intelligence while building the applications [14,

15]. The face validation API provided in Microsoft Cognitive

Services offers the ability to control the likelihood that people

identified in different photos are the same person. This API returns

a confidence score about the likelihood of two faces belonging to

a single person. There are a total of 30,000 process limits and a

maximum of 20 process limits per minute [16]. The

implementation was developed using the Microsoft Cognitive face

verification API, which includes the ability to detect, identify,

analyse, organize, and tag faces in photos.

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Computer Eng. Dept., Selcuk University, Konya – 42002, TURKEY 2 Inst. of Natural Sciences, Selcuk University, Konya – 42002, TURKEY

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 294–298 | 295

Genymotion is an Android emulator with a set of features for

interacting with a virtual Android environment. With Genymotion,

Android applications can be developed, run and tested with a wide

variety of virtual devices [17]. The app was tested by creating an

Android device in the Genymotion environment.

In this study, Android application in Java language was developed

in Android Studio 3.0.1 environment.

The working principle of the developed application is described

below:

1. The captured image from the camera or the selected image from

the gallery is uploaded to the application as the 1st photo

2. The system detects the faces in the 1st photo after the 1st photo

is uploaded.

3. The faces detected in the photo are listed on the side of the photo.

4. The captured image from the camera or the selected image from

the gallery is uploaded to the application, 1st photo

5. The system detects the faces in the 2nd photo after the 2nd photo

is uploaded.

6. The faces detected in the photo are listed on the side of the photo.

7. Make a selection from the recognized faces in photo 1 and photo

2 and click on the "Verify" button.

8. The system detects whether the selected two faces are the same

person and notifies the user.

9. The similarity rate of the selected faces is shown as a value

between 0 and 100.

3. Experimental Results

The developed application consists of two parts. These sections are

the section on application and about. In the application section,

there is a button that directs the division, which allows the face

identification and verification process to be performed. The About

section contains information about the application. Fig. 1 shows a

screenshot of the application area.

Fig. 1. Application Start Page

When you click on the Start button, the section that performs the

face identification and verification process opens. In Fig. 2, this

section has been given a screenshot. In this section, two images

should be selected for identification and verification. There is a

button showing the records of the transactions made.

Fig. 2. Image Selection Phase

In the realization of the face recognition process, two sample

photographs were used. These photographs are shown in Fig. 3.

In the application, click the select image button and the 1st photo

shown in Fig. 3 is selected. The system identifies the faces and lists

the detected faces in thumbnail. The list contains the ability to

scroll down. In addition, the number of faces detected in the first

photo is given. A visual representation of this application is given

in Fig. 4.

Fig. 3. Sample Photographs

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 294–298 | 296

Fig. 4. Face Identification and Verification Phase – Image 0

After image selection process is done as Image 0, image selection

process is done as Image 1. For this, click the Select Image button

in the lower part. The second photo shown in Fig. 3 is selected and

the identified faces are listed in thumbnail. The number of faces

detected in the bottom section is given to the user. A visual

representation of this application is given in Fig. 5.

Fig. 5. Face Identification and Verification Phase – Image 1

After selecting images as Image 0 and Image 1, the faces to be

compared are selected from the listed face images beside both

images. When the Verify button is clicked, the user is informed

numerically whether the selected two faces belong to the same

person. In Fig. 5, two face selections were made and the result

shown in Fig. 6 is obtained when the confirm button is clicked. In

Fig. 6, it is seen that there are different people because the

similarity rate of the two faces selected is 17%. To show the result

of the validation process being performed on the same person, the

selected face from Image 0 has been replaced with a different

person face from Image 1. When the validation process is applied,

it is given that the faces selected in Image 0 and Image 1 belong to

the same person because the similarity ratio is 71%. A screenshot

of this application is shown in Fig. 7.

Fig. 6. Performing Verification on Faces of Different Persons

Fig. 7. Performing Verification on Faces of Same Person

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 294–298 | 297

Records of all transactions made are kept. A list of all the actions

performed when clicking the Show Record button is visible. In this

section where the records are displayed, information about resizing

of two images uploaded to the system, successful detection of the

images, the number of faces detected in the visuals, request for

sending the faces to be verified to the server, the status of the

returned response and the result of the response appear. Fig. 8

shows a screenshot of the system register with two

requests. The first response was successful and the result was

negative. The second response was successful and the result was

positive.

In practice, each process takes about 3 seconds, such as uploading

photos to the system, detecting and verifying faces in uploaded

photos because Microsoft Cognitive Services is used in the process

of performing transactions. An information screen informing the

user is developed as shown in Fig. 9.

Fig. 8. Screenshot of the system register

Fig. 9. User Information Screens

The determination of whether the system is stable or not has been

done on photographs containing more than one person. As shown

in Fig. 10, a test for face detection and verification of persons

labelled A, B, C, D, E was performed. This test involves comparing

each face in the photo with each other.

(a)

(b)

Fig. 10. Face detection and verification of persons

The result of this test is given in Table 1. The rows represent the

faces on the upper photo in Fig. 10 and the columns represent the

faces on the lower. Similarity ratios of 25% comparisons were

obtained. When the results obtained were observed, it was seen that

all persons were correctly verified.

Table 1. Similarity Rates after Verification Process

Faces in Fig. 10 (a)

A B C D E

Fac

es i

n F

ig. 10

(b

)

A

87 13 46 32 31

B

08 87 19 35 16

C

27 24 82 39 32

D

35 38 35 80 27

E

24 13 29 38 85

When the results obtained were observed, the highest similarity

rate was obtained as 87% for persons A and B. Despite being the

same person in both photographs, the similarity rate for D person

was 80%. Some of the reasons for the similarity ratio being

different are the effect of the person's posing position and light.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 294–298 | 298

4. Conclusion

The Cognitive Services are very up - and - coming, assuring and

sum of these services are already being used in various application

areas. The dev elopers use these services to improve the

applications by adding some new, rich and intelligent features.

These services basically aim at providing more personal and rich

computing experiences for users as well as developers. The range

of application areas of these cognitive services is less but it can be

expanded by using their services along with other modern services

and techniques.

In this work, an Android application was developed that includes

face identification and authentication methods that can work on

mobile devices. The developed application provides a security

measure through its face identification and authentication feature.

This application can be integrated into various applications and

used in systems such as user login. User authentication can be

performed by scanning the user's face without the need for

password entry.

References

[1] F. C. Elbizim, M. C. Kasapbasi, (2017) Implementation and

Evaluation of Face Recognition Based Identification System,

International Journal of Intelligent Systems and Applications in

Engineering, pp. 17-20

[2] A. Eleyan, (2017) Simple and Novel Approach for Image

Representation with Application to Face Recognition, International

Journal of Intelligent Systems and Applications in Engineering, vol.

5(3), pp. 89-93.

[3] S. Nikan, M. Ahmadi, (2015) Performance Evaluation of Different

Feature Extractors and Classifiers for Recognition of Human Faces

with Low Resolution Images, International Journal of Intelligent

Systems and Applications in Engineering, vol. 3(2), pp. 72-77.

[4] W. Zhao, R. Chellappa, P. J. Phillips and A. Rosenfeld, 2003, “Face

Recognition: A Literature Survey”, ACM Computing. Surveys;

35(4): 399–458

[5] J. A. Lee, A. Szabo, Y. Li, (2013) AIR: An Agent for Robust Image

Matching and Retrieval, International Journal of Intelligent Systems

and Applications in Engineering, vol. 1(2), pp. 34-39.

[6] H. Mliki, E. Fendri, M. Hammami, 2015, “Face Recognition

Through Different Facial Expressions”, Journal of Signal Processing

Systems, vol 81(3), pp. 433-446.

[7] N. Delbiaggio, A comparison of facial recognition’s algorithms,

Bachelor's Degree, Haaga-Helia, 2017

[8] E. Sütçüler, 2006, “Real-Time Face Localization and Recognition

System by Using Video Sequences”, Msc Thesis, Yıldız Technical

University.

[9] P. Kirci, G. Kurt, (2016) Long Term and Remote Health Monitoring

with Smart Phones, International Journal of Intelligent Systems and

Applications in Engineering, vol. 4(4), pp. 80-83.

[10] G. Eryigit, G. Celikkaya, (2017) Use of NLP Techniques for an

Enhanced Mobile Personal Assistant: The Case of Turkish,

International Journal of Intelligent Systems and Applications in

Engineering, vol. 5(3), pp. 94-104.

[11] Beginning Android Programming with Android Studio, John Wiley

& Sons, 2016, ISBN: 1118707427, 9781118707425

[12] R. Arora, J. Sharma, U. Mali, A. Sharma, p. Raina (2018), Microsoft

Cognitive Services, International Journal of Engineering Science

and Computing, vol. 8(4), pp. 17323-17326.

[13] A. Del Sole, Introducing Microsoft Cognitive Services, In:

Microsoft Computer Vision APIs Distilled. Apress, Berkeley, CA,

https://doi.org/10.1007/978-1-4842-3342-9_1

[14] A.A. Vinay, S.S. Vinay, J.B. Rituparna, A. Tushar, K.N.

Balasubramanya Murthya, S. Natarajanb. (2015), Cloud Based Big

Data Analytics Framework for Face Recognition in Social Networks

using Machine Learning, 2nd International Symposium on Big Data

and Cloud Computing (ISBCC’15) .

[15] L. Larsen (2017), Learning Microsoft Cognitive Services - Create

intelligent apps with vision, speech, language, and knowledge

capabilities using Microsoft Cognitive Services, Packt Publishing

Ltd.

[16] Face Verification, https://azure.microsoft.com/tr-

tr/services/cognitive-services/face [Access Date: 05.01.2018]

[17] Genymotion User Guide – Version 2.11

https://docs.genymotion.com/pdf/PDF_User_Guide/Genymotion-

2.11-User-Guide.pdf [Access Date: 06.01.2018]

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 299–305 | 299

Psychological Stress Detection from Social Media Data using a Novel

Hybrid Model

Mohammed Mahmood Ali*,1, Shaikha Hajera2

Accepted : 08/10/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: One of the mental threat for individual’s health identified is Psychological stress from social media data. Hence, necessity is to

predict and manage stress before it turns into a serious problem. However, Conventional stress detection methods exist, that rely on

psychological scales & physiological devices that need full of victims participation which is time-consuming, complex and expensive.

With the trending growth of social networks, people are addicted towards sharing personal moods via social media platforms to influence

other users, leading to stressfulness. The developed novel hybrid model Psychological Stress Detection (PSD), automatically detect the

victims’s psychological stress from social media data. It comprises of three (3) modules Probabilistic Naïve Bayes Classifier, Visual

(Hue, Saturation, Value) and Social, to leverage text, image post and social interaction information; we defined set of stress-related

textual ‘F = {f1, f2, f3, f4}’, visual ‘vF = {vf1, vf2}’, and social features ‘sf’ to predict stress from social media content. Experimental

results show the proposed PSD model improves the detection process when compared to TensiStrength and Teenchat framework, PSD

achieves 95% of Precision rate. PSD model will assist in developing stress detection tools for mental health agencies.

Keywords: Psychological Stress Detection(PSD); Social Media platforms; Naïve Bayes, Visual, Social, mental health agencies;;

1. Introduction

Psychological stress is considered the biggest threat to an

individual’s health. Three of every hundred people in

metropolitan areas are assessed to suffer from stress. Long-lasting

stress may lead to many severe physical and mental problems,

such as depressions, insomnia, and even suicide. Stress and

suicide are closely interlinked. At its worst, anxiety can lead to

suicide [1]. According to World Health Organization (WHO)

over 56 million Indian’s suffer from depression and Corporate

sector in India as well reported an increase in stress over the last

two years, a survey by workplace solutions provider Regus in

2015, reported 57% of corporate India is under stress and about

one in five people in the country need counseling, either

psychological or psychiatric. The National Institute of Mental

Health and Neurosciences (NIMHANS) published a mental

health assessment said 5% of the population suffers from

depression as of 2016. Thus, recognizing stress or depression at

an early stage is critical for reducing suicidal deaths and

deliberate self-harm across the spectrum. A range of sensors and

non-textual methods have been developed to detect stress as

sensors can offer data regarding the intensity and quality of a

person’s internal affect Experience [7]. For examples, H.

Kurniawan presents stress detection from speech and galvanic

skin response signals [3], and [6].S. Greene describes the stress

detection method utilizing sensor which monitors Heart Rate

Variability (HRV) [3], [5], and [7]. These sensors are expensive

and most of the Traditional stress detection methods are mainly

based on one-on-one interviews conducted by psychologists or

self-report questionnaires [13] or wearable sensors[16]. However,

these are usually labor & time-consuming and are reactive

methods. The query is, are there any appropriate and practical

methods for stress recognition?

The rapid growth of Social Media is Changing People’s Life, as

Well as Research in Healthcare and Wellness. Report from

Global Social Media research 2018 [4] shows there are in total

3.196 billion active social media users worldwide and among

Fig 1: Sample Post containing text, image and social interactions (i.e.,

comments and likes)

them Facebook is the most popular social network with total 2.01

billion active users monthly and Twitter is the fastest growing

social networks with total of 328 million active users. With the

quick advancement of social networking sites, individuals are

excited to use it as a platform to express their moods and day-to-

day life actions making it a popular platform to express thoughts

via posts. People share thoughts, express emotions, record daily

habits via textual and image posts and also interconnect with

friends, a sample is depicted in Figure 1. As these social media

data reveal user’s reallife situations and emotions in a timely

fashion, it offers new prospects for assessing, modeling, and

mining user’s activity patterns through social networks, such

social data can find its theoretical basis in psychology research

and Encoding emotional information in text is a common practice

especially in online interactions [2]. Thus, we can obtain

linguistic and visual content from individual’s posts that may

indicate stress related symptoms that makes the detection of

user’s psychological stress feasible through their posts, posting

behavior and social interaction on micro-blog or social media.

The rest of paper is organized as follows: The Section 2 gives an

overview of the related work and also describes the problem

statement. Section 3 introduces the definitions of the proposed

features and presents the hybrid model Psychological Stress

Detection (PSD) for stress recognition and describes the

algorithmic structure and modules of the proposed PSD model.

Section 4 presents experimental results. Lastly, Section 5

provides the conclusion and discusses future scope.

I am feeling very sad and don’t feel like doing

anything. :( depressed.

I hate my manager and job schedule, very stressed.

1Associate Professor, Osmania University, Telangana -500034, INDIA.

* Corresponding Author: Email: [email protected]. 2 Research Scholar, MJCET,OU, Telangana -500034, INDIA.

Alex Russo 28 Feb 2018

Comments (0) Likes (9)

Alex Russo 28 Feb 2018

Likes (9) Comments (0)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 299–305 | 300

2. Related Work And Problem Defination

There have been research efforts on harnessing social media data

for developing mental and physical healthcare tools. K. Lee

et.al’s [8] proposed to leverage social media data for real-time

disease investigation; while [10] tried to link the vocabulary gaps

between health seekers and providers using the community

generated health data. Psychological stress detection is

interrelated to the topics of sentiment assessment and emotion

recognition. Our recent work [25] presented relative analysis of

various psychological stress recognition Approaches. Study on

Emotion Detection in Social Networks, Computer-aided

recognition, scrutiny, and application of emotion, especially in

social media, have drawn considerable attention in current years

[11], [12], and [15]. Associations between psychological stress

and personality traits can be a thought-provoking issue to

contemplate [17], [18], and [20]. Several studies on social

networks based emotion analysis uses text-based linguistic

features and typical classification approaches. As discussed

above a range of sensors and non-textual methods have been

developed to detect stress but they are expensive. With the

development of social networks people are willing to share their

daily events and moods, and interact with friends through the

social networks, making it possible to leverage online social

network data for stress detection. There are also some exploration

[22], and [27] using user posting contents on social network to

identify user’s stress revealed that leveraging this social media

data for healthcare, and in particular stress detection, is feasible.

However, these mechanisms mainly leverage the textual contents

and consider only single post of an individual in social networks.

In certainty, data in social networks are typically composed of

chronological and inter-connected items from diverse sources i.e.,

it also contains images posts apart from textual posts, making it

be actually multi-media data and single post reflects the instant

emotion but people’s emotion or psychological stress states are

habitually more persistent, fluctuating over different time periods.

In recent years, extensive study started to focus on emotion

detection in social networks from sequential post series [21],

[23], and [24]. Our work is to leverage sequential post (multi-

media data) content over specific sampling period to detect stress.

The proposed novel hybrid model Psychological Stress Detection

(PSD) defines set of stress-related textual ‘F = {f1, f2, f3, f4}’,

visual ‘vF = {vf1, vf2}’, and social ‘sf’ features that makes the

detection efficient and effective leveraging textual and visual post

content.

Before presenting our problem definition, let’s first define

necessary notations. Let P be a set of users posts on a social

media network i.e., pi ∈ P. Let U be set of users. Each user ui ∈

U, posts on social media containing text, image. Where i denotes

number of posts and users that can range from 1 to ‘n’ i.e., i = {1,

2, 3, …, n}.

Stress state: stress state denoted by‘s’ of user ui ∈ U at

time t is represented as a triple (s, ui, t), or . In the study,

a binary stress state ∈ {0, 1} is considered, where = 1

specifies that user ui is stressed at time t, and = 0

specifies that the user is non-stressed at time t.

Stress feature set: Let F, vF and sf be the stress feature sets

in study i.e., F = {f1, f2, f3, f4}, vF ={vf1,vf2}, sf , where

f1 – f4 are linguistic stress lexicon features, vf1 and vf2 are

visual features and sf is social feature as shown above in

table 4.2.(page number 19).

Problem Statement: To Detect Psychological Stress state

‘s’ Given a series of user posts P. To address this problem

we first define the stress feature sets they are Textual

feature (F), Visual feature (vF) and social feature (sf) then

propose Psychological Stress Detection (PSD) model to

leverage the social media content for stress detection.

3. Proposed Psychological Stress Detection (PSD)

Model

3.1. Feature Categorizations

To cope with the problem of psychological stress detection we

have defined set of stress-related textual ‘F = {f1, f2, f3, f4}’,

visual ‘vF = {vf1, vf2}’, social ‘sf’ features. Table 1 shows

categorization of these features. Textual Feature is a set of four

features, F = {f1, f2, f3, f4}, these feature stores word count of a

particular sentence comparing them with linguistic stress lexicon

explained in Table 1. ‘f1’ stores number of stress and negative

words present in the sentence posted. ‘f2’ stores number of

positive emotion words present in the sentence posted, ‘f3’ stores

negative emoticons if present and ‘f4’ stores negating words

present in the sentence posted. These features help the

probabilistic module in calculating the probability of each class

and labeling the textual post to its particular class as shown in

Figure 4.

Visual Feature is a set of two features, vF = {vf1, vf2} they stores

the mean value of the brightness and saturation, which is

calculated in the visual module algorithm as shown in Figure 5.

Value of Image brightness and saturation will be compared with

threshold value to categorize the particular image post as stress or

non-stress as shown in Figure 3. Social Feature ‘sf’ stores the

mean value of number of likes a particular post gets it give us the

social attention degree that means when user posts are negative or

stressful it gets more attention from the friends of that particular

user via likes or comments. Thus, we will find the mean of

number of likes and categories the posts accordingly as shown in

Figure 3.

The detail Architecture of the model is shown in the Figure 2.

Psychological stress detection has many challenges to overcome

them first have categorized features as discussed in section 3.1,

now will design modules to utilize these features for stress

detection. Probabilistic module fetches textual feature ‘F = {f1,

f2, f3, f4}’, from content of user’s post comparing with

Linguistic Stress Lexicon. Visual module fetches visual features

‘vF = {vf1, vf2}’ from image post, and Social module will fetch

social attention feature ‘sf’ as discussed in section 3.1.

3.2. Architecture

Figure 2 shows the architecture of our model. There are three

types of information that we can use as the initial inputs, i.e.,

Textual features, Visual features and Social features whose

detailed computation will be described later. We address the

solution through the following three key modules:

Design the Probabilistic module to fetch textual feature ‘F

= {f1, f2, f3, f4}’, from content of user’s post comparing

with Linguistic Stress Lexicon.

Visual module fetches visual features ‘vF = {vf1, vf2}’

from image post, and

Social module fetches social attention feature ‘sf’ as

discussed in section 3.1.

Steps involved in the architecture of the proposed system are as

follows:

i. The first step captures the Post and messages sent

between the users of Social Networking Site (SNS).

These messages are saved in the database.

ii. In this step, a text post is taken and this text message is

converted to plain text removing stop-words (such as

articles, preposition). Next, will check if the user has

also posted an image if yes, then will send it to visual

module to process its mean value of brightness and

saturation. Next, will check social attention feature to

calculate the mean of likes.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 299–305 | 301

Table 1: Illustrates set of knowledge based pre-defined logical rules.

Short name Sample Words/ Images to be detected Feature

notation Description

Textual feature (F): Rule 1 – Algorithm 1 (Figure 3) will check threshold values to tag the textual posts as stress

Stress Lexicon

stress,depress,anxious,sad,depress,anger,pain,restless,sleepless

,failure,sorrow,lonely,quit,worry,nervous,mental,break,anguish

,argu,arrogant,atrocious,bereave,brokenhearted,Hurt, ugly,

anger, annoy, boredom, sadness, grief, disapproval, low self-

esteem, callous, despair, devastate, hate……..

f1 Number of stress words and negative

words.

Positive

emotion

Lexicon

Love, joy, trust, surprise, care, happy, amazement, interest….. f2 Number of positive words.

Negative

Emoticons

"%-(", ")-:", "):", ")o:" , ":(", ":*(", ":,(", ":-", ":-&", ":-

("……. f3

Number of popular Negative

emoticons.

Negating words

Lexicon

Cannot, cant, couldn't, didn’t, doesn't, don’t, hasn't, isn't,

never, no, not, shouldn't, won't, wouldn't. f4 Number of negating words.

Visual feature (vF): Rule 2 - Algorithm 1 (Figure 3) will check threshold values to tag Visual posts as stress

Brightness vf1 Mean of image pixels i.e., brightness

Saturation vf2 Mean of saturation.

Social feature(sf): Rule 3 - Algorithm 1 (Figure 3) will check threshold values

Social attention sf Number of likes and mean value of it.

iii. This is the main step in which will compare the plain

text with linguistic stress lexicon to find out the number

of stress words, negative words present in the text and

store the result into feature set denoted by ‘f1-f4’. Next,

will extract the image features and store the result into

feature set denoted by ‘vf1’ and ‘vf2’. Lastly will

calculate the number of likes and stored it in feature

‘sf’ as shown in Table 1.

iv. The Textual (linguistic) features are fed into the

probabilistic module to predict the class of the text

post. The module uses maximum likelihood estimates

for the detection. The visual features are fed into the

visual module, where the brightness and saturation of

the image is calculated as shown in figure 5 and

compared with the threshold value as shown Figure 3.

Social feature is fed into social module to mean of likes

and to compare it with a threshold value shown in

Figure 3.

v. In the last step, results are displayed to the user if the

user posts satisfies all the threshold values then only

his/her posts will be tagged as stress as shown in Fig. 3

and if the user is status is stressed then will send the

recommendation to the user regarding stress

management.

1.1.1. Probabilistic Module

To fetch textual feature from the posts will make use of Naïve

Bayes classifier which is based on Bayes theorem. This classifier

algorithm uses conditional independence, means it assumes that

an attribute value on a given class is independent of the values of

other attributes. It is simple yet efficient classifier that is used to

classify uncertain data. It works on the principles of Bayes

Probability theorem that means it will be having a set of known

results for some words and it compares those results with the

words in a particular sentence.

Fig.2: Psychological Stress Detection (PSD) Architecture.

The Bayes theorem is as follows: Let X = { } be a

set of ‘n’ attributes. In Bayesian, ‘X’ is considered as evidence

and ‘H’ is some hypothesis means, the data of X belongs to

specific class ‘C’. We have to determine , the probability

Likes (9)

Social networking site / Instant Messenger

Feature extraction

program Post of User

Social

feature

Textual

feature

Visual

feature

Program

to evaluate

one week

posts

User Stress label

Text, image captured

Probabilist

-ic module

Visual

module Social

module

i

ii

iii

iv

v

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 299–305 | 302

that the hypothesis H holds given evidence i.e. data sample X

[30]. According to Bayes theorem the is expressed as

shown in equation 1 or 2.

or

Where is any one of the class in the list i.e., ‘No-Stress’,

‘Stress’ and Neutral. - are Number of words in the given

sentence, here wordn = fn, where n = 4 refer Table 1. If textual

feature fetched from the given sentence matches the words in the

linguistic stress lexicon explained in Table 1 will assign class to

that sentence according to the probability score of each class.

The detail computation of probabilistic module is shown in

Probabilistic Module Algorithm Figure 4. As discussed in section

3.1 the textual Feature are a set of four features, F = {f1, f2, f3,

f4}, these feature stores word count of a particular sentence

comparing them with linguistic stress lexicon explained in Table

1. ‘f1’ stores number of stress and negative words present in the

sentence posted. ‘f2’ stores number of positive emotion words

present in the sentence posted, ‘f3’ stores negative emoticons if

present and ‘f4’ stores negating words present in the sentence

posted. These features help the probability module in labeling the

posted sentence is having stress or non-stress by calculating the

probability of each class.

3.2.1. Visual Module

Based on previous work on color psychology theories [28], we

combine the following features as the visual middle-level

representation: - Saturation: the mean value of saturation and its

contrast. It describes the colorfulness and the differences of an

image. Psychological experiments in S. R. Ireland’s [29] find out

that people under stress and anxiety prefer lower saturation than

normal states, revealing the correlation between stress and

saturation of images. Brightness: the mean value of brightness

and its contrast. It illustrates the perception elicited by the

luminance and the related differences of an image (e.g., low

brightness makes people feel negative, while high brightness

elicits mainly positive emotional associations). To calculate

image brightness first, we need to (briefly) analyze what is the

result of the average value of the sum of the RGB channels. For

humans, it is meaningless. Is pink brighter than green ? I.e., why

would you want (0, 255, 0) to give a lower brightness value than

(255, 0, 255) ? Also, is a mid gray (128, 128, 128) bright just like

a mid green (128, 255, 0) ? To take this into consideration, I only

deal with the color brightness of the channel as is done in the

HSV color space [14]. This is simply the maximum value of a

given RGB triplet. The rest is heuristics. Let max_rgb = max

(RGB-i) for some point i. If max_rgb is lower than 128

(assuming a 8bpp image), then we found a new point i that is

dark, otherwise it is light. Doing this for every point i, we

get A points that are light and B points that are dark. If (A - B)/(A

+ B) >= 0.5 then we say the image is light. Note that if every

point is dark, then you get a value of -1, conversely if every point

is light you get +1. The previous formula can be tweaked so you

can accept images barely dark. In the code I named the variable

as fuzzy, but it does no justice to the fuzzy field in Image

Processing. So, we say the image is light if (A - B)/(A + B) +

fuzzy >= 0.5.Following is the mathematical relationship between

RGB space to HSI (hue, saturation, and intensity) or Hue,

Saturation, Value (HSV) color space model.

• Brightness or Intensity formula

• Saturation formula

Where vf1 is brightness or Intensity and vf2 is saturation as

shown in equation 3 and 4. The detail computation of visual

module is shown in Visual Module Algorithm figure 5. Visual

Feature is a set of two features, vF = {vf1, vf2} they stores the

mean value of the brightness and saturation, which is calculated

in the visual module algorithm. Value of Image brightness and

saturation will be compared with threshold value to categorize the

particular image post as stress or non-stress as shown in Figure 3.

3.2.2. Social Module

Besides the text content and image content of a post, some

additional features such as likes can also imply one’s stress state

to some degree. We define a post’s social attention degree based

on these additional features into social attributes. As apparently

stressful posts may attract more attention from friends. The

number of comments, likes reveals how much attention a post

attracts. To find out the changes of attention degree of one’s post,

we first calculate the sample mean sf shown in equation 5, all M -

values and total number of items in the samples.

Social Feature ‘sf’ stores the mean value of number of likes a

particular post gets it give us the social attention degree that

means when user posts are negative or stressful it gets more

attention from the friends of that particular user via likes or

comments. Thus, we will find the mean of number of likes and

categories the posts accordingly as shown in Figure 3.

Algorithm 1: Psychological stress Detector (PSD) Algorithm

Require: pi, ui, (Social media posts P of users U).

Ensure: ‘s’ (Detection of Stress state i.e., stress or non-stress)

1 Initialize P, U, s = {stress , non-stress}, F ,vF ,sf .

2 u = getusername() //getusername function returns

current user name

3 While(i < U and P != NULL ) do

4 F = Stress (P) //stress function of probability module calculates

stress from textual posts and return stress state

5 vF = getImageLightness (image) //getImageLightness function of visual module calculates and return image brightness

6 sf = num_likes() //num_likes function of social module calculates

and return number of likes and mean value of the likes.

7 if(F = 1 and vF <= 0.5 or sf >= 5) then //To display stress status

for a week.

8 stress_status = stress //tag that user having

stress

9 //if the user have more than five posts tagged as stress in a week then system will display result as stress.

if(loggedin_user == u and stress_status >= 5 ) then

10 s = stress //stress state

11 else

12 s = non-stress //stress state

13 end if

14 end if

15 end while

Fig.3: Psychological Stress Detection (PSD) Algorithm.

(1)

(2)

(3)

(4)

(5)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 299–305 | 303

Algorithm 2: Probabilistic Module Algorithm

Require: sentence (Textual post passed by PSD algorithm)

Ensure: stress_class (textual post belongs to which class i.e.,

stress(x), non-stress(y) and neutral(z)).

1 Function Stress (sentence )

//given training data D which consists of database of words belonging to different class i.e., stress (x), non-stress (y) and

neutral(z)

2 Initialize stress_class = {x,y}

//Implication that the analyzed text has 1/3 chance of being in either

of the 3 categories 3 x = 0.333

4 y =0.333

5 z =0.334

//find total number of word frequency of each class

6 n = total number of word frequency

//find conditional probability of words occurrence given a class

7 P(fm / x) = wordcount / n(x)

8 P(fm/y) = wordcount / n(y)

9 P(fm/z) = wordcount/n(z)

//classify a new sentence based on probability

10 P(sentence / stress_class) = product-of[p(x) * P(fm/ stress-class)]

//assign sentence to class that has higher probability

11 If ( stress_class == x )then

12 return 1

13 else

14 return 0

15 end if

16 End

Fig.4: Probabilistic Module Algorithm.

Algorithm 3:Visual Module Algorithm

Require: ‘y’ (visual post passed by PSD algorithm)

Ensure: ‘vF’ (brightness and saturation value)

1 Function getImageLightness (y)

2 Initialize imageData = y.getImageData(0,0,width,height)

3 Initialize data = imageData

4 Initialize r,g,b,vf1,vf2,avg

5 for(x=0 , len=data.length; x < len ; x+=4)do

6 r = data[x]

7 g = data[x+1]

8 b = data[x+2]

9 vf1 = (r+g+b)/3

10 vf2 = 1-[(3/(r+g+b))*(min(r,g,b))]

11 avg += vf1+vf2

12 end for

13 return avg/255

14 End

Fig.5: Visual Module Algorithm.

4. Experimental Results

The proposed PSD model is compared with Mike Thelwall’s

TensiStrength [27] which uses a lexical approach with lists of

terms related to stress and non-stress. TensiStrength’s terms are

not only synonyms for stress, anxiety and frustration but also

terms related to anger and negative emotions because stress can

be a response to negative events and can cause negative

emotions. It also attempts to detect the opposite state to stress,

non-stress, through a parallel approach. This method is mainly

based on text data in the social media, whereas other

correspondingly significant content, like images and social

actions are ignored and it is capable of detecting stress only from

a single textual post, while single post reflects the instant emotion

expressed, people’s emotion or psychological stress states are

usually more persistent and changes over different time periods

and social media allows users to express their mental thoughts or

feelings through not only textual posts but also in visual post i.e.,

through images.

Fig.6: Comparison of Accuracy and F1-score % based on different parameters.

The proposed PSD is designed as such it takes into consideration

textual and multi-media (images) data as well to detect stress.

The sample scenario assumes at-most two posts (texts plus emoji

plus image) per day and results are analyzed we achieved 83.3%

of accuracy and 88.8% of F1-score. To further analyze the model

we have taken more scenarios. In the Scenario2 we have

considered post including emojis i.e., Text + Emoji and analyzed

posts for a week and achieved 91.6% of accuracy and 92.2% of

F1-score. In scenario3 we have considered post which includes

emojis and image along with text post i.e., Text + Emoji + image

shown in Table 2 and based on these these parameters we have

achieved highest F1-score of 94% and accuracy of 91.9%. For

further test our model, we have considered data from our

application’s database which is real time content and Table 3

shows the details of that data. The collected data consists of 1,920

posts based on these data we have evaluated PSD model and it

showed an increased accuracy compared to J. Huang’s Teenchat

[22] and M.thelwall’s TensiStrength[27] system for stress

detection. The figure 7 shows a comparison of efficiency of the

PSD model from the result shown in table 4 we see that PSD

model achieved better detection performance and showed F1-

Score of 95.7% whereas TensiStrength shows 5.9% of less F1-

score i.e., 90.10% and J. Huang’s Teenchat [22] shows 15.5% of

less F1-score i.e., 80.20% and when compared to our PSD model.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 299–305 | 304

Table 2: Shows a scenario of user’s posts and result analysis (Text + Emoji +Image).

S.no. Posts Bayes Score Image I-score Class

1 My life is finished very alone without my parents sad feeling, in

pain.

0.727

0.252 S

2 I am in pain, can’t take it anymore hate life depressed. 0.8

0.237 S

3 Its sad when someone you know becomes you knew depressed to

lose my friend 0.7

0.116 S

4 I feel tense, anxious or have nervous indigestion. 0.72

0.419 S

5 I can't do this anymore. My body is physically incapable to take

this pain 0.67

0.165 S

6 I am so disturbed. Life is hard. :( , %-( 0.666

0.419 S

7 This pain is never ending hate this feeling sad, dishearten. 0.728

0.376 S

8 I am separated from my husband, i am alone sad and depress. 0.67

0.331 S

9 My sister came from USA happy to see her. 0.25

0.651 NS

10 I got a best singer award happy to receive this honor 0.1

0.7 NS

11 I lost my job dont have money feels demotivated cant take this

mental pressure. 0.444

0.5 NS

Analysis

Total posts to be detected as stressed 9

Number of posts detected as stressed 8

Precision 99.9%

Recall 88.8%

Accuracy 91.9%

F1-score 94%

Table 4: Comparison of efficiency & effectiveness using different model

Models Text

Text +

Emoji

Text + Emoji

+ Image Accuracy F1-Score CPU time

Teenchat [28] 0.784 0.802 28s

TensiStrength [39] 0.891 0.901 180s

PSD model 0.93 0.957 35s

Table 3: Details of Data taken for result analysis.

Details Count

Number of posts 1920

Number of Users 20

Number of weeks 8

Posts per week 12

Precision 95%

parameters, as shown Tensistrength [27] achieved 89% of

accuracy and it supports only text parameter for stress detection

and for the same. J. Huang’s Teenchat [22] model achieved

91.5% of accuracy based on all three parameters but the time

taken by CPU is much higher than our PSD model which

decreases its efficiency, our PSD model achieved 93% of

accuracy when all parameters are considered and also take very

less CPU time to process and detect stress from social media

posts of individual users.

Fig.7: Comparison of efficiency of the models based on Accuracy

and F1-score.

Parameter

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 299–305 | 305

5. Conclusion

Psychological Stress is an increasing problem in the world, it is

essential to monitor and control it regularly. In this paper, we

presented a Model for detecting user’s stress status from user’s

daily social media data, leveraging post’s content along with

user’s social interactions. The inefficiency of the conventional

methods [5], [16], [19], [22], and [27] motivates to adopt a novel

technique which is transparent and inexpensive. Similar to our

another ontological approach is developed for stress detection

that works better on Textual stress words [31]. Thus, the

proposed strategy utilizes not only social media data (text and

images), but also uses social interaction content to detect

psychological stress employing a Novel Hybrid model i.e.,

Psychological Stress Detection (PSD). To completely leverage

user’s posts content, we proposed a hybrid PSD model which

combines the probabilistic module (Bayes classifier) with the

visual module (HSV model). In our model the user’s stress states

are monitored daily, weekly, bi-weekly or monthly so that it

improves the accuracy of stress detection at different levels.

Experimental results show that the proposed model can improve

the detection performance when compared to J. Huang’s

Teenchat [22], and TensiStrength [27], and achieved 95% of

precision and 93% of accuracy.

5.1 Future Scope

Recent studies show apart from leveraging cross-media content

and social interactions, Social Connectivity also plays a vital role

in detecting stress states of an individual, which is a useful

reference for future related studies. Some more useful references

for future related studies:

Need to explore our research on stressor responsible for

stress as it is noteworthy to know the type of stress the

user is dealing.

Support for short form and slang language used in text

post to be included.

Support for Multilingual languages to be included.

References

[1] J. Herbert, “Fortnightly review: Stress, the brain, and mental

illness”, British Medical J., pp. 530–535, Vol. 315, No. 7107, 1997.

[2] Liew, and Jasy Suet Yan, “fine-grained emotion detection in

microblog text”, Dissertations – ALL. pp 440, 2016. [3] F.-T. Sun, C. Kuo, H.-T. Cheng, S. Buthpitiya, P. Collins, and M.

L. Griss, “Activity-Aware Mental Stress Detection Using

Physiological Sensors”, In Proc. Of Intl. Conf. on Mobile Computing, Application, and Services (MobiCASE), Santa Clara,

CA, 2010.

[4] Social,https://www.smartinsights.com/social-media-marketing/social-media-strategy/new-global-social-media-

research.(online).

[5] H. Kurniawan, A.V. Maslov, and M. Pechenizkiy, “Stress detection from speech and galvanic skin response signals”, in:

Proceedings of the 26th IEEE International Symposium on

Computer-Based Medical Systems, pp. 209–214, 2013. [6] V.P. Patil, K.K. Nayak, and M. Saxena, “Voice stress detection”,

Int. J. Electr. Electron. Comput. Eng., pp. 148–154, 2013.

[7] S. Greene, H. Thapliyal, and A. C. Holt, “A survey of affective computing for stress detection: Evaluating technologies in stress

detection for better health”, IEEE Consum. Electron. Mag., Vol. 5, No. 4, pp. 44–56, 2016.

[8] K. Lee, A. Agrawal, and A. Choudhary, “Real-time disease

surveillance using twitter data: Demonstration on FLU and cancer”, in Proc. ACM SIGKDD Int. Conf. Knowl. Discovery

Data Mining, pp. 1474–1477, 2013.

[9] Sztajzel, “Heart rate variability: a noninvasive electrocardiographic method to measure the autonomic nervous

system”, Swiss Med. Weekly, pp. 514–522, 2004.

[10] L. Nie, Y.-L. Zhao, M. Akbari, J. Shen, and T.-S. Chua, “Bridging

the vocabulary gap between health seekers and healthcare

knowledge,” IEEE Trans. Knowl. Data Eng., vol. 27, no. 2, pp.

396–409, Feb. 2015.

[11] G. Coppersmith, C. Harman, and M. Dredze, “Measuring post

traumatic stress disorder in twitter,” in Proc. Int. Conf. Weblogs

Soc. Media, 2014, pp. 579–582.

[12] R. Fan, J. Zhao, Y. Chen, and K. Xu, “Anger is more influential

than joy: Sentiment correlation in weibo,” PLoS One, vol. 9, 2014,

Art. no. e110184.

[13] Stressassessments,https://www.nysut.org/~/media/files/nysut/resou

rces/2013/april/socialservices/socialservices_stressassessments.pdf?la=e.(online).

[14] HSV,

https://shodhganga.inflibnet.ac.in/bitstream/10603/18613/12/12_chapert-3 (online).

[15] Y. Zhang, J. Tang, J. Sun, Y. Chen, and J. Rao, “Moodcast:

Emotion prediction via dynamic continuous factor graph model,”

Proc. IEEE 13th Int. Conf. Data Mining, 2010, pp. 1193–1198.

[16] A. Sano, and R. W. Picard, “Stress recognition using wearable sensors and mobile phones”, In Proceedings of ACII, pp. 671–676,

2013.

[17] G. Farnadi, et al., “Computational personality recognition in social

media,” UserModel User-Adapted Interaction, vol. 26, pp. 109–

142, 2016.

[18] J. Golbeck, C. Robles, M. Edmondson, and K. Turner, “Predicting

personality from Twitter,” in Proc. IEEE 3rd Int. Conf. Privacy,

Security, Risk Trust, IEEE 3rd Int. Conf. Soc. Comput., 2011, pp.

149–156.

[19] A. Fernandes, R. Helawar, R. Lokesh, T. Tari, and A. V. Shahapurkar, “Determination of stress using blood pressure and

galvanic skin response”, In Proceedings of the International

Conference on Communication and Network Technologies (ICCNT’14), pp. 165–168, Sivakasi, India, 2014.

[20] B. Verhoeven, W. Daelemans, and B. Plank, “Twisty: A

multilingual twitter stylometry corpus for gender and personality

profiling,” in Proc. 10th Int. Conf. Language Resources Eval., PP.

1632–1637 2016.

[21] F. A. Pozzi, D. Maccagnola, E. Fersini, and E. Messina, “Enhance

user-level sentiment analysis on microblogs with approval

relations,” in Proc. 13th Int. Conf. AI* IA: Advances Artif. Intell.,

PP. 133–144, 2013.

[22] J. Huang, Q. Li, Y. Xue, T. Cheng, S. Xu, J. Jia, and L. Feng.:

“Teenchat: a chatterbot system for sensing and releasing

adolescents’ stress”, In: X. Yin, K. Ho, D. Zeng, U. Aickelin, R. Zhou, H. Wang. (eds.) HIS LNCS, Vol. 9085, pp. 133–145.

Springer, Heidelberg, 2015.

[23] C. Tan, L. Lee, J. Tang, L. Jiang, M. Zhou, and P. Li, “User-level

sentiment analysis incorporating social networks,” in Proc.

SIGKDD Int. Conf. Knowl. Discovery Data Mining, 2011, pp.

1397–1405.

[24] J.W. Pennebaker, R.J. Booth, and M.E. Francis, “Linguistic

Inquiry and Word Count (LIWC)”, 2007. [25] Shaikha Hajera and Mohammed Mahmood Ali, “A Comparative

Analysis of Psychological Stress Detection Methods”, In IJCEM,

Vol. 21 Issue 2, March 2018. [26] Y.R. Tausczik, and J. W. Pennebaker, “The psychological

meaning of words: LIWC and computerized text analysis

methods”, Journal of Language and Social Psychology, pp. 24-54, 2010.

[27] Thelwall M. “TensiStrength: stress and relaxation magnitude

detection for social media texts”, J Inf Process Manag pp: 106–121, 2017.

[28] X. Wang, J. Jia, H. Liao, and L. Cai, “Affective image

colorization”, Journal of Computer Science and Technology, Vol. 27, No. 6, pp. 1119-1128, 2012.

[29] S. R. Ireland, Y. M. Warren, and L. G. Herringer, “Anxiety and

color saturation preference”, Perceptual and Motor Skills, Vol. 75, pp. 545-546, 1992.

[30] Keogh, Eamonn, “Naive bayes classifier” Accessed: Sept. 1

(2006): 2018. [31] Mohammed Mahmood Ali, Mohd Tajuddin, and M Kabee, “SDF:

psychological Stress Detection Framework from Microblogs using

Predefined rules and Ontologies”, International Journal of Intelligent Systems and Applications in Engineering, Vol. 6, issue

2, pp. 158-164, 2018.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 306–310 | 306

A Modified Artificial Algae Algorithm for Large Scale Global

Optimization Problems

Havva Gul KOCER1, Sait Ali UYMAZ2

Accepted : 18/12/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: Optimization technology is used to accelerate decision-making processes and to increase the quality of decision making in

management and engineering problems. The development technology has made real world problems large and complex. Many optimization

methods that proposed for solving large-scale global optimization (LSGO) problems suffer from the “curse of dimensionality”, which

implies that their performance deteriorates quickly as the dimensionality of the search space increases. Therefore, more efficient and robust

algorithms are needed. When literature on large-scale optimization problems is examined, it is seen that algorithms with effective global

search ability have better results. For the purpose, in this paper Modified Artificial Algae Algorithm (MAAA) is proposed by modifying

original version of Artificial Algae Algorithm (AAA) inspiring by Differential Evolution Algorithm (DE)’s mutation strategies. AAA and

MAAA are compared with each other by operating with the first 10 benchmark functions of CEC2010 Special Session on Large Scale

Global Optimization. The results show that hybridization process that applied by updating an additional fourth dimension with mutation

strategies of DE after the helical motion of the AAA algorithm, contributes exploration phase and improves the AAA performance on

LSGO.

Keywords: artificial algae algorithm,CEC2010 benchmark,large scale global optimization

1. Introduction

There are various high dimensional problems in the modern world.

Because most modern data, such as computational biology, data

from consumer preferences, images, videos, are often high

dimensional. Optimization is a technique for solving problems and

it is the calculation of the parameters that will produce the best

result of a function. However, as the dimension of problem

increases, the process of reaching a definitive solution becomes

very complex, difficult and costly. This is because when the

dimensionality increases, the volume of the space increases

exponentially and probability of finding good solutions with

limited search amount decreases dramatically. R. Bellman used

"curse of dimensionality" term in the introduction of his book

“Dynamic programming” in 1957 for this situation [1]. This term

refers to the difficulties of finding optimum in high-dimensional

space using extensive search. Therefore, there is a need to use

effective and efficient new techniques beyond what exists to solve

large-scale optimization problems.

Metaheuristic methods that find the most suitable solution around

the real solution at an acceptable time, are preferred as an effective

solution technique for large-scale optimization problems. While

metaheuristic algorithms are shown excellent search abilities when

applied to some small dimensional problems, usually their

performance deteriorates quickly as the dimensionality of search

space increases. Therefore, a more efficient search strategy is

required to explore all the promising regions in a given time or

fitness evaluation count limits.

Some attempts have been made by the researchers, such as modify

original version of an algorithm [2,3], hybridize with another

algorithm [4], using additional local search methods [5,6] or

develop solution techniques [7] to make metaheuristic algorithms

suitable to overcome the difficulties of solving large-scale

optimization problems.

Inspired by Differential Evolution mutation strategies, this paper

presents an improved variant of the AAA algorithm called

Modified Artificial Algae Algorithm(MAAA) for solving large-

scale global optimization problems.

The rest of the paper is organized as follows. First, Artificial Algae

Algorithm, Differential Evolution are described briefly. Then

Modified Artificial Algae Algorithm is described. In the following

section benchmark set used in tests, parameters of algorithms and

working environment are analyzed. Finally, results and

interpretations are given and are followed by conclusions and

recommendations.

2. Artificial Algae Algorithm

Artificial Algae Algorithm (AAA) is a bio-inspired meta-heuristic

optimization algorithm designed to solve continuous and real-

value optimization problems. It is developed by inspiration of

foraging behaviour of micro-algae. In this population-based

algorithm each individual is named as artificial algal colony and

each artificial algal colony corresponds to a solution in the problem

space. Artificial algae, like algae in real life, moves to the light

source (better solutions) in order to do photosynthesis. Their

movement is in the form of helical swimming. Their life cycle

consists of adapting to the environment, changing the dominant

species and multiplying by mitosis. AAA mimics these lifecycle

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Distance Education Application and Research Center, Selçuk University,

Konya – 42002, TURKEY 2 Department of Computer Engineering, Konya Technical University,

Konya – 42002, TURKEY

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 306–310 | 307

behaviours in three phases. These phases are helical movement,

evolutionary process and adaptation. Pseudo-code of AAA is given

in Table 1 [8,9].

AAA is proposed in 2015 by Uymaz et al. [8]. In the same year

Uymaz et al. made a modification on AAA by multi light source

[9]. And then many researchers have been worked on AAA. For

example, Babalık et.al. studied AAA for multi-objective

optimization [10], Zhang et al. studied binary version of AAA [11],

Kumar & Dhillon studied hybrid version of AAA by hybridized

with simplex search method (SSM) [12]. AAA applied on real-

world problems such as wind turbine placement problem [13],

economic load dispatch problem [12]. But there is no study on

large-scale optimization with AAA. The lack of a study of AAA

on a large-scale optimization is the motivation of this work. Results

of the original version of AAA were promising, so a modification

was applied on AAA to increase the performance on LSGO.

Table 1. AAA pseudo-code

1 : generate an initial population of n algal colonies with random

solutions

2 : evaluate f(xi), i = 1, 2, …, D

3 : while stopping condition not reached do

4 : for i = 1 to n do

5 : while energy of ith colony not finished do

6 : modify the colony with helical movement

7 : end while

8 : end for

9 : apply evolutionary strategy

10 : apply adaptation strategy

11 : end while

3. Differential Evolution

Differential evolution (DE), proposed by Storn and Price [14], is

an efficient and versatile population-based direct search algorithm.

Among its advantages are its simple structure, ease of use, speed,

and robustness, which enables its application on many real-world

applications.

DE is an effective algorithm frequently used in the LSGO

[15,16,17]. The algorithm achieved third rank in CEC 2008 special

session and competition on high-dimensional real-parameter

optimization [18]. Therefore, effects of crossover and mutation

operators of DE algorithm [19,20] for LSGO performance is also

investigated.

Mutation is treated as a random change of some parameter. Thanks

to the changes in the mutation phase, the solution point

representing the chromosome is moved in the solution space. In

DE-literature, a parent vector from the current generation is called

target vector, a mutant vector obtained through the differential

mutation operation is known as donor vector and finally an

offspring formed by recombining the donor with the target vector

is called trial vector [20]. In mutation stage, several DE trial vector

generation strategies have been proposed. Well-known mutation

strategies were listed in [21,22]. One of the DE mutation strategies

in the literature defined as follows:

𝑉𝑔+1 = 𝑥𝑔 + 𝐹(𝑥𝑏𝑔 − 𝑥𝑔) + 𝐹(𝑥𝑟1𝑔 − 𝑥𝑟2𝑔) + 𝐹(𝑥𝑟3𝑔 − 𝑥𝑟4𝑔) (1)

In (1) which calculates the donor vector V, the index g + 1

indicates the next generation and the index b refers to the best

individual. In this equitation, four distinct parameter vectors r1, r2,

r3, r4 are sampled randomly from the current population. The

difference of these vectors is scaled by a scalar number F and the

scaled difference is added to the grand total.

4. Modified Artificial Algae Algorithm

In order to achieve successful results in the LSGO field, it is

necessary to use an algorithm which has powerful exploration and

exploitation. While the evolutionary and adaptation processes of

AAA support exploitation mechanism, the process of helical

movement contributes to both exploration and exploitation

mechanisms. In other words, exploration and exploitation process

of AAA are provided by a spiral movement through modification

of algal colonies. Therefore, in order to increase the search

capabilities, we made the modification in the process of helical

movement. Helical movement pattern like as follow Fig. 1 [23].

Fig. 1. Helical movement pattern of Algae.

Helical movement is a process which updates each algal colony. In

this process, colony is selected with tournament selection and only

three algal cells of each algal colony, which are also selected

randomly and represented as k, l and m, are modified by the

following formulas:

Calculation for Linear Movement:

𝑥𝑖𝑚𝑡+1 = 𝑥𝑖𝑚

𝑡 + (𝑥𝑗𝑚𝑡 − 𝑥𝑖𝑚

𝑡 )(∆ − 𝜏𝑡(𝑥𝑖))𝑝 (2)

Calculation for Angular Movement:

𝑥𝑖𝑘𝑡+1 = 𝑥𝑖𝑘

𝑡 + (𝑥𝑗𝑘𝑡 − 𝑥𝑖𝑘

𝑡 )(∆ − 𝜏𝑡(𝑥𝑖)) cos 𝛼 (3)

𝑥𝑖𝑙𝑡+1 = 𝑥𝑖𝑙

𝑡 + (𝑥𝑗𝑙𝑡 − 𝑥𝑖𝑙

𝑡 )(∆ − 𝜏𝑡(𝑥𝑖)) sin 𝛽 (4)

Parameters which are used in (2), (3), (4) , are ∆ and 𝜏 , shear

force and friction surface of algal colonies, respectively.

It was thought that making an update in a fourth algal cell of each

algal colony immediately after the helical movement in AAA could

be useful for increasing exploration capability of the algorithm.

Therefore, a modification has been made in AAA inspired by

Differential Evolution mutation strategy in (1) and this approach

was called as MAAA. While DE uses differences of random

vectors in current population, we used difference of helical

movement calculation results(𝑥𝑖𝑚𝑡+1, 𝑥𝑖𝑘

𝑡+1, 𝑥𝑖𝑙𝑡+1). Random selected

fourth algal cell (n) of same algal colony are modified by following

formula:

𝑥𝑖𝑛𝑡+1 = 𝑥𝑏𝑒𝑠𝑡

𝑡 + (𝑥𝑖𝑚𝑡+1 − 𝑥𝑖𝑘

𝑡+1) + (𝑥𝑖𝑘𝑡+1 − 𝑥𝑖𝑙

𝑡+1) + (𝑥𝑖𝑙𝑡+1 − 𝑥𝑖𝑚

𝑡+1) (5)

Pseudo-code of MAAA is given in Table 2.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 306–310 | 308

Table 2. Pseudo-code of modification in AAA helical movement phase

for every algal colony (second and fourth steps)

1 : Select another colony with tournament selection

2 : Select four algal cells (k, l, m and n) in the colony randomly

3 : Modification the algal cells(k, l and m) of the colony with (2), (3) and (4)

4 : Modification the fourth cell (n) of the same colony with (5)

which uses the results of (2), (3), (4).

5 : Decrease energy caused by movement

6 : If new solution is better, move new position else decrease

energy by metabolism

7 : If energy of the colony did not finish go to step 1

8 : If the colony did not find better solution increase starvation of the colony

5. Performance Evaluation

5.1. Experimental Environment

The computer platform used to perform the experiments was based

on an Intel(R) Xeon(R) CPU E5-2650 2.00 GHz processor, 8.70

GB of RAM, and the Microsoft Windows Server 2012 operating

system. All experimental works were conducted by using Matlab

(Release R2010a). Matlab code v.1. of Artificial Algae Algorithm

is released in [24]

5.2. Benchmark Set

Since 2005, competitions have been organized in the LSGO field

traditionally at CEC (IEEE Evolutionary Computing Congress)

private sessions. Common rules and benchmark sets were

presented in 2008, 2010 and 2013 to evaluate algorithms fairly in

these competitions. In this paper, we used 10 functions of CEC

2010 benchmark set given in Table 3. All functions are the

minimization problems and the optimum function values are zero

for all the problems. Detailed information about benchmark set is

provided in [25].

Table 3. Used benchmark function set of CEC2010 LSGO

No Name Range

F1 Shifted Elliptic Function [−100, 100]

F2 Shifted Rastrigin [−5, 5]

F3 Shifted Ackley [−32, 32] F4 Single-group Shifted and m-rotated Elliptic [−100, 100]

F5 Single-group Shifted and m-rotated Rastrigin [−5, 5]

F6 Single-group Shifted and m-rotated Ackley [−32, 32] F7 Single-group Shifted m-dimensional Schwefel’s

Problem 1.2

[−100, 100]

F8 Single-group Shifted m-dimensional Rosenbrock [−100, 100] F9 D/2m -group Shifted and m-rotated Elliptic [−100, 100]

F10 D/2m -group Shifted and m-rotated Rastrigin [−5, 5]

5.3. Parameter Settings

For this study, the recommendations in [8] were followed by

setting the energy loss, e = 0.3, the shear force, Δ = 2, the

adaptation parameter, Ap = 0.5 and the population size, N=40.

5.4. Testing Procedure

All reported results in this study were the averages over 25

simulations. In ordered results, the results of the 1st (Best), 13th

(Median), and 25th (Worst) trial, as well as the trial mean (Mean),

and standard deviation (Std.) are presented.

Dimension of problems in the benchmark set are 1000. Maximum

fitness evaluation (MaxFEs) was set to 3e6.

6. Results and Discussion

The original version of AAA and the proposed modified version of

AAA, called MAAA, are run on the benchmark set in Table 3, in

equal conditions and same parameters. Table 5 shows the results

of AAA and MAAA. For remarking the best algorithm, we have

put minimum values for each function in bold.

According to Table 5, the proposed MAAA algorithm obtained

better results in seven functions of ten than AAA algorithm in both

average and best values. In F1 and F2 functions, MAAA showed a

significant improvement over AAA. Proposed MAAA has also

reached the optimum solution in F1. MAAA and AAA have similar

standard deviations.

(a) F1

(b) F7

(c) F8

Fig. 2. Convergence curves of sample functions for AAA and MAAA

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 306–310 | 309

The convergence curves of the AAA and MAAA for the following

selected three problems: F1, F7, F8 are presented in Fig. 2. Y axis

represents mean values of 25 independent runs of related function.

X axis represents a recorded point in each 3000 fitness evaluations.

For example, 200 value at x axis corresponds to 6e5 FEs.

Fig.2a shows that MAAA much faster than AAA until 6e5 FEs.

After that point, AAA continues to find better results but MAAA

draws a constant graph until 1.5e6 FEs. After that point, MAAA

convergence to optimum very fast.

Fig.2b shows that MAAA has same convergence characteristic

with AAA and therefore they draw parallel convergence graphic.

But MAAA slightly faster so find better results than AAA.

Fig.2c shows that MAAA much faster than AAA until around 8e5

FEs. However, after that point, AAA pass MAAA with very close

performance.

Table 4. Summary Information of Compared Algorithms

Algorithm Description

SDENS [26] Sequential Differential Evolution enhanced by

neighborhood search

DNSPSO [27]

A hybrid PSO algorithm which employs a

diversity enhancing mechanism and

neighborhood search strategies

DASA [28] Differential Ant-Stigmergy Algorithm is an

Ant-colony optimization-based algorithm.

DECC-G [7]

Differential Evolution Cooperative

Co-evolutionary with random grouping and adaptive weighting

LOCUST SWARMS ( LS )

[29]

A multi-optima search technique explicitly designed for non-globally convex search

spaces.

Table 6 shows the test results of MAAA and other methods given

in Table 4. Comparisons are drawn from the mean of 25

independent runs at 3.0e+6 FEs. MAAA results are obtained from

running own codes. Results of other LSGO algorithms are directly

taken from references which mentioned in header of the Table 6

columns.

While MAAA obtains best results in F1, F3 and F8, DASA obtains

best results in F2, F4, F7 and F9. Therefore, MAAA is the second

algorithm after DASA. Ranks part of Table 6 shows that the order

of each algorithm among all algorithms. Average ranks of each

algorithm proves that MAAA is the best second algorithm among

six algorithms.

7. Conclusion

Metaheuristic optimization algorithms often show sufficient search

capabilities on small optimization problems, but when applied to

large and complex problems with more than a hundred decision

variables, often lose their efficacy and performance [23].

Therefore, there is a need to develop effective and efficient new

modifications or techniques. For this purpose, we made a

modification on original version of AAA and investigated effects

on LSGO performance.

The obtained results show that additional dimension modification

after helical movement of AAA algorithm by inspired from DE

algorithm mutation strategies let the AAA algorithm perform

better in large dimension problems.

For future works, the performance of AAA can be improved by

applying LSGO solutions techniques or hybridizing with another

powerful optimization algorithm.

Table 5. Test results of AAA and MAAA

AAA MAAA

Best Median Mean Std. Best Median Mean Std.

F1 4.39E-17 2.60E-16 3.02E-16 2.42E-16 0.00E+00 2.71E-28 3.06E-26 1.24E-25

F2 2.98E+00 6.26E+00 8.14E+00 5.29E+00 6.08E+01 8.08E+01 8.06E+01 1.45E+01

F3 2.89E-11 8.99E-11 5.48E-03 2.51E-02 2.52E-13 4.12E-13 4.71E-13 2.06E-13

F4 1.23E+13 1.89E+13 1.86E+13 4.35E+12 8.22E+12 1.55E+13 1.62E+13 4.62E+12

F5 3.34E+08 4.21E+08 4.32E+08 5.73E+07 2.97E+08 4.19E+08 4.24E+08 6.88E+07

F6 1.89E+07 1.96E+07 1.95E+07 1.90E+05 1.78E+07 1.95E+07 1.94E+07 3.76E+05

F7 7.55E+09 1.24E+10 1.28E+10 3.61E+09 4.26E+09 9.48E+09 9.72E+09 2.87E+09

F8 4.15E+03 1.61E+06 2.99E+06 3.88E+06 1.46E+04 1.90E+06 9.38E+06 1.31E+07

F9 1.65E+08 2.21E+08 2.13E+08 2.51E+07 1.49E+08 1.86E+08 1.91E+08 2.16E+07

F10 5.65E+03 6.16E+03 6.12E+03 2.64E+02 6.41E+03 6.75E+03 6.78E+03 3.02E+02

Table 6. Comparison with other algorithms

Mean Values Ranks

MAAA SDENS[26] DNSPSO[27] DASA[28] DECC-G[6] LS[29] MAAA SDENS DNSPSO DASA DECC-G LS

F1 3.06E-26 5.73E-06 1.87E+07 1.52E-21 2.93E-07 3.48E+05 1 4 6 2 3 5

F2 8.06E+01 2.21E+03 5.85E+03 8.48E+00 1.31E+03 1.24E+03 2 5 6 1 4 3

F3 4.71E-13 2.70E-05 1.93E+01 7.20E-11 1.39E+00 2.67E+00 1 3 6 2 4 5

F4 1.62E+13 5.11E+12 2.25E+12 5.05E+11 1.70E+13 6.65E+12 5 3 2 1 6 4

F5 4.24E+08 1.18E+08 1.57E+08 6.20E+08 2.63E+08 3.60E+08 5 1 2 6 3 4

F6 1.94E+07 2.02E-04 1.75E+06 1.97E+07 4.96E+06 1.91E+07 5 1 2 6 3 4

F7 9.72E+09 1.20E+08 8.60E+06 7.78E+00 1.63E+08 6.32E+08 6 3 2 1 4 5

F8 9.38E+06 5.12E+07 1.31E+07 4.98E+07 6.44E+07 3.10E+07 1 5 2 4 6 3

F9 1.91E+08 5.63E+08 3.16E+08 3.60E+07 3.21E+08 2.22E+08 2 6 4 1 5 3

F10 6.78E+03 6.87E+03 6.90E+03 7.29E+03 1.06E+04 6.04E+03 2 3 4 5 6 1

Average Rank = 3.0 3.4 3.6 2.9 4.4 3.7

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 306–310 | 310

Acknowledgment

This study was financially supported by The Coordinatorship of

Scientific Research Projects of Selçuk University [Grant no:

18101012].

References

[1] R. E. Bellman, Dynamic Programming, Princeton University Press,

Princeton, 1957.

[2] Y. Sun, X. Wang, Y. Chen and Z. Liu,” A modified whale optimization

algorithm for large-scale global optimization problems”, Expert

Systems with Applications, 2018.

[3] W. Long, J. Jiao, X. Liang and M. Tang,” Inspired grey wolf optimizer

for solving large-scale function optimization problems”, Applied

Mathematical Modelling, 2018, pp. 112-126.

[4] A. F. Ali, M. A. Tawhid, “A hybrid particle swarm optimization and

genetic algorithm with population partitioning for large scale

optimization problems”, Ain Shams Engineering Journal, 2017.

[5] N. Noman and H. Iba, “Enhancing differential evolution performance

with local search for high dimensional function optimization”,

In Proceedings of the 7th annual conference on Genetic and

evolutionary computation, 2005, pp. 967-974.

[6] D. Molina, M. Lozano and F. Herrera,”MA-SW-Chains: Memetic

algorithm based on local search chains for large scale continuous

global optimization”, In Evolutionary Computation (CEC), 2010, pp.

1-8.

[7] Z. Yang, K. Tang, and X. Yao,” Large scale evolutionary optimization

using cooperative coevolution”. Information Sciences, 2008, pp. 2985-

2999.

[8] S.A. Uymaz, G. Tezel and E. Yel,” Artificial algae algorithm (AAA)

for nonlinear global optimization”, Applied Soft Computing, 2015a,

pp. 153–171.

[9] S.A. Uymaz, G. Tezel and E. Yel, “Artificial algae algorithm with

multi-light source for numerical optimization and applications”,

Biosystems, 2015b, pp. 25-38.

[10] A.Babalik, A. Ozkis, S.A Uymaz and M.S. Kiran, ”A multi-objective

artificial algae algorithm”, Applied Soft Computing, 2018, pp. 377-

395.

[11] X. Zhang, C. Wu, J. Li, X. Wang, Z. Yang, J. M. Lee and K. H. Jung,

”Binary artificial algae algorithm for multidimensional knapsack

problems”, Applied Soft Computing, 2016, 43, pp. 583-595.

[12] M. Kumar and J. S. Dhillon, “Hybrid artificial algae algorithm for

economic load dispatch”, Applied Soft Computing, 2018, 71, pp.89-

109.

[13] M. Beşkirli, İ. Koç, H. Haklı and H. Kodaz,” A new optimization

algorithm for solving wind turbine placement problem: Binary

artificial algae algorithm”, Renewable Energy, 2018, 121, pp. 301-308.

[14] R. Storn and K. V. Price, “Differential evolution: A simple and

efficient adaptive scheme for global optimization over continuous

spaces,” ICSI, USA, Tech. Rep. TR-95-012, 1995 [Online]. Available:

http://www1.icsi.berkeley.edu/~storn/TR-95-012.pdf

[15] Y. Gao and Y. Wang, “A memetic differential evolutionary algorithm

for high dimensional function spaces optimization”, in Proc. 3rd ICNC

20, 2007, pp. 188–192.

[16] Z. Yang, K. Tang, and X. Yao, “Differential evolution for high

dimensional function optimization,” in Proc. IEEE Congr. Evol.

Comput., 2007, pp. 3523–3530.

[17] J. Brest, A. Zamuda, B. Boskovic, M. S. Maucec and V. Zumer,

“High dimensional real-parameter optimization using self-adaptive

Differential evolution algorithm with population size reduction,” in

Proc. IEEE Congr. Evol. Comput., 2008, pp. 2032–2039

[18] K. Tang, X. Yao, P. N. Suganthan, C. MacNish, Y. P. Chen, C. M.

Chen, and Z. Yang, “Benchmark functions for the CEC’2008 special

session and competition on large scale global optimization”, Nature

Inspired Comput. Applicat. Lab., USTC, China, Nanyang Technol.

Univ., Singapore, Tech. Rep., 2007

[19] T. Keskintürk, “Diferansiyel Gelişim Algoritması”, İstanbul Ticaret

Üniversitesi Fen Bilimleri Dergisi Yıl: 5 Sayı: 9 Bahar 2006, pp.85-99

[20] S. Das and P. N. Suganthan,”Differential evolution: a survey of the

state-of-the-art”. IEEE transactions on evolutionary computation,

2011, pp 4-31.

[21] W. Gong, Á. Fialho, Z. Cai and H. Li,” Adaptive strategy selection in

differential evolution for numerical optimization: an empirical study”,

Information Sciences, 2011, 181(24), pp. 5364-5386.

[22] A. Banitalebi, M. I. A. Aziz and Z. A. Aziz,” A self-adaptive binary

differential evolution algorithm for large scale binary optimization

problems”, Information Sciences, 2016, 367, pp.487-511.

[23] H. C. Lund and J. W. G. Lund, “Freshwater Algae-Their micoscopic

world explored”, Biopress Limited, Bristol, England, 1996.

[24] http://aaa.selcuk.edu.tr/, Accessed on: October 8, 2018.

[25] K. Tang, X. Li, P. N. Suganthan, Z. Yang, T. Weise,” Benchmark

functions for the CEC’2010 special session and competition on large

scale global”, Nature Inspired Comput. Applicat. Lab., Tech. Rep.,

2009

[26] H. Wang, Z. Wu, S. Rahnamayan and D. Jiang,” Sequential DE

enhanced by neighborhood search for large scale global optimization”,

in Proc. IEEE Congr. Evol. Comput., 2010, pp. 1-7.

[27] H. Wang, H. Sun, C. Li and S. Rahnamayan and J.S. Pan,” Diversity

enhanced particle swarm optimization with neighborhood search”

Information Sciences, 2013, pp. 119-135.

[28] P. Korošec, K. Tashkova and J. Šilc,” The differential ant-stigmergy

algorithm for large-scale global optimization”, in Proc. IEEE Congr.

Evol. Comput, 2010, pp. 1-8.

[29] S. Chen, ”Locust Swarms for Large Scale Global Optimization of

Nonseparable Problems”, Kukkonen, Benchmarking the Classic

Differential Evolution Algorithm on Large-Scale Global Optimization

Google Scholar.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 311

An Investigation of the Effect of Meteorological Parameters on Wind

Speed Estimation Using Machine Learning Algorithms

Cem Emeksiz*1, Gulden Demir2

Accepted : 16/12/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: Wind speed is the most important parameter of the wind energy conversion system. Therefore temperature, humiditiy and

pressure data, which has significant effect on the wind speed, have become extremely important. In the literature, various models have

been used to realize the wind speed estimation. In this study; Six different data mining algorithms were used to determine the effect of

meteorological parameters on wind speed estimation. The data were collected from the measurement station established on the campus of

Gaziosmanpaşa University. We focused on the bagging algorithm to determine the appropriate combination of wind speed estimates. The

bagging algorithm was used for the first time in estimation of wind speed by taking into account meteorological parameters. To find the

most efficiency method on such problem 10-fold cross validation technique was used for comparision. From results, It is concluded that

bagging algorithm and temperature-humiditiy-pressure combination showed the best performance. Additionaly, temperature and pressure

data are more effective in the wind speed estimation.

Keywords: Bagging algorithm, Data mining, Renewable energy, Wind speed estimation.

1. Introduction

With industrialization, basic energy resources emerge as an

indispensable phenomenon to maintain the daily life of human

beings. In industrial societies, fossil fuels are used as the main

energy source. Nowadays, while a large part of energy needs are

met by fossil fuels, it is a necessity to turn to alternative sources

when the negative impact of these resources on the environment is

taken into consideration. At the same time, global warming,

seasonal changes and the political approaches of the countries have

accelerated the search for alternative energy sources in the future

where fossil fuel reserves can not meet their needs. This is

especially the case for wind energy and solar energy, which are

among the renewable energy sources.

Renewable energy sources can be renewed owing to the variety of

energy that does not emit to the environment, which is thought to

take its power from the sun and never to be consumed. When we

look at the energy policies of developed and developing countries,

we see that they are turning to renewable energy sources and

accelerating their efforts to develop these resources. Solar energy,

wind energy, hydrolic energy, biomass energy, hydrogen energy,

geothermal energy, wave energy, tidal energy are the categories of

renewable energy resources.

According to the 2017 Renewable Energy Global Situation Report,

in 2016, a total of 161 GW renewable energy is installed in the

power system. The total global capacity has increased by about 9%

compared to 2015 and is almost 2.017 GW at the end of the year.

In 2016, about 47% of the newly established renewable energy

capacity was solar energy. This ratio was followed by wind energy

with 34% and hydroelectric energy with 15.5% [1].

Wind energy is at the forefront of the renewable energy sources

that are the most used owing to cost efficiency in the world. In the

past, wind energy was used to move sails, run ships, windmills and

in irrigation. Recently, the use of wind turbines operating in the

world by wind energy in the production of electricity has become

increasingly widespread. According to the 2016 Global Wind

Statistics report on the situation of the global wind industry of

Global Wind Energy Council (GWEC), countries with the highest

installed power are shown in Table 1 [2].

Table1 The top 10 cumulative capacity DEC 2016

Country MW %Share

PR China 168.690 34.7

USA 82.184 16.9

Germany 50.018 10.3

India 28.700 5.9

Spain 23.074 4.7

United K. 14.543 3.0

France 12.066 2.5

Canada 11.900 2.4

Brazil 10.740 2.2

Italy 9.257 1.9

Rest of theworld 75.577 15.5

Total TOP 10 411.172 84

World Total 486.749 100

To benefit technologically from wind energy; it is very important

to know the possibilities of utilization, to determine the zones with

high wind energy potential, to be able to predict the wind

characteristics and wind speeds. In particular, estimating wind

speeds is necessary for estimating the energy expected to be

generated from wind turbines in short, medium and long period.

According to these estimation values, the profitability of the power

generation plants can be calculated and it is determined whether or

not the investment of wind energy in a region will be profitable. In

this way, the operating and production costs can be calculated more

accurately [3].

Wind energy estimations are generally produced by hybrid models

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Electric-Electronic Eng. Tokat Gaziosmanpasa University, Tokat, Turkey 2 Electric-Electronic Eng. Tokat Gaziosmanpasa University, Tokat, Turkey

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 312

that are used physically, statistically, or both. The physical method

is generally used for long-term estimates, while statistical methods

are used for short-term estimates. In the physical approach, maps

of the plant area to be used in the models, topographic maps and

smoothness maps are prepared using Geographic Information

System (GIS) software. The plant settlement is done on maps, and

the flow of wind is modeled by computational fluid dynamics

software. In the statistical estimation models, the relationship

between the local winds and the numerical weather forecast results

is established by using the meteorological data in the past. The

statistical approach uses models such as Neural networks (NN),

Support Vector Machine (SVM), Fuzzy Logic and Regression

Trees.

When studies on wind speed estimations are examined; Monthly

mean temperature measured from Arak weather station from 1960

to 2005, dew point in sunny hours, relative humidity, average wind

speed, saturated vapor pressure data and monthly potential

evaporation temperature estimation were studied. Bagging model

was found to be functionally more successful as a result of Root

Mean Square Error (RMSE), Mean Squared Error (MSE) and

Correlation Coefficient (CC) values [4]. Cadenas et al. used an

univariate Autoregressive Integrated Moving Average (ARIMA)

model and a multivariate Nonlinear Autoregressive (NARX)

model to estimate the wind speed. The data set consists of the

parameters taken from the stations in two different regions. Mean

Absolute Error (MAE) and Mean Square Error (MSE) values are

used to analyze the results. Using the NARX model has proved to

be more accurate. It has also been proposed to incorporate

additional meteorological parameters in wind speed prediction

models [5].

To estimate the wind speed in India, data from 31 provinces with

differentiated geographical conditions from National Aeronautics

and Space Administration (NASA) and the Generalized

Regression Neural Networks (GRNN) model were used. 26

provincial data were used as training data and the other 5 provincial

data were used as test data. Based on CC and MSE values, GRNN

model was found to be very efficient for predicting long-term wind

speed [6]. Zeng et.al. studied the effect of different sampling

frequencies on short-term wind speed determination and power

estimation with support vector machines [7]. Khandelwal et.al.

have developed a time series forecasting model that combines

discrete wavelet transforms with ARIMA and artificial neural

network model [8]. Khanna et.al. studied the determination of time

series properties in wind power generation [9]. For wind turbines

in Korea, wind energy was estimated with ANFIS, Sequential

Minimal Optimization (SMO), K-Nearest Neighbor (KNN) and

Artificial Neural Networks (ANN) models using hourly and daily

wind speed, wind direction, temperature and time intervals. ANFIS

and SMO models have been shown to perform well in predicting

wind energy [10]. Wanga et.al. used copula theory to estimate wind

speed. They have determined that representative wind data can be

obtained much more reasonably with the conditional distribution

[11]. Velo et.al have presented a method for determining the

annual average wind speed in a complex land area using neural

networks, where only short term data are available. As neural

network inputs, they used wind speed and direction data obtained

from a single station [12]. Timur et.al. estimated the wind speed

using LinearRegression, K-Nearest Neighbor (KNN), Bagging,

DecisionTable, and REPTree classification algorithms for Istanbul

Göztepe region. In the cross-validation option, 5 and 10 values

were given to k in the cross validation option, and the Bagging

algorithm was observed to be more successful in terms of the CC

and RMSE (Root Mean Square Error) values of the most

successful result [13]. In a study Zontul et al. Between 2001 and

2007, wind speed estimation with WEKA was performed by using

cross-validation method with yearly, monthly, daily wind direction

data were obtained from meteorological data of Kırklareli

Province. The correlation coefficient between the true value and

the estimated value was found to be a successful classification

method of the Bagging model with 0.8154 [14]. Based on the

Wavelet Packet Transformation (WPD), the Cross-Optimization

(CSO) Algorithm and the Artificial Neural Networks, Meng et.al.

have developed a new hybrid model to predict a short-term wind

speed of 1 hour intervals up to a 5-hour period with two different

wind speed ranges in the wind observation station in Rotterdam

[15].

In this study, the wind speeds, which are accepted as the most

important inputs of wind energy, were estimated by using machine

learning algorithms. Six different algorithms were used in the

estimation and Bagging algorithm realized the best estimation with

the lowest error and highest correlation coefficient (CC) among

these algorithms. The meteorological parameters that affect the

estimation of wind speed in the study were also examined and

results indicate that the combination of temperature-humidity-

pressure combination realized a prediction with lower error rate.

2. MATERIALS AND METHODS

In this study; the data mining algorithms used for wind speed

estimation are Bagging Algorithm, SMOreg Algorithm, K-Star

Algorithm, Multilayer Perceptron, REP Tree Algorithm and M5P

algorithm.

2.1. Bagging Algorithm

Bagging algorithm is a method that is used to improve results are

obtained from machine learning classification algorithms. Leo

Breiman formulated the Bagging algorithm in 1994. This

algorithm is an abbreviation of "bootstrap aggregating" [16]. The

generalization capability of machine learning algorithms can be

evolved by using the Bagging algorithm. The bagging algorithm is

analyzing in a shorter time using parallel learning as an alternative

to algorithms that have long analysis time, such as artificial neural

networks. The process of the Bagging algorithm is shown in Figure

1 [17]. Sample sets (Si) are created by using random return

selection. The corresponding classifier Ci is trained based on the

training set Si, respectively. In this way, the final pattern

recognition can be obtained by using the primary recognition

results.

Fig. 1. Bagging algorithm description

The bagging method creates a sequence of classifiers Ci, i=1,…,M

in respect to modifications of the training set. These classifiers are

combined into a compound classifier. The prediction of the

compound classifier is given as a weighted combination of

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 313

individual classifier predictions [18]:

𝐶(𝑑𝑚) = 𝑠𝑖𝑔𝑛(∑ 𝛼𝑖𝐶𝑖(𝑑𝑚)𝑀𝑖=1 ) (1)

The meaning of the above given formula can be interpreted as a

voting procedure. An example dm is classified to the class for

which the majority of particular classifiers vote. Parameters αi,

i=1,…,M are determined in such way that more precise classifiers

have stronger influence on the final prediction than less precise

classifiers. Bagging decrease the variance of your single estimate

so the result may be a model with higher stability.

2.2. SMOreg Algorithm

SMOreg Algorithm is used with Support Vector Machine (SVM)

that is classification method that divides the data into two

categories. There is a training data set for classification process.

This training data belongs to any of the two categories. After that

SVM estimates that new instance is in which category. It is aimed

to create a hyperplane in a n-dimensional space in this process.

Whereby two datasets can be separated which data is nearest to

hyper plane is named support vector. Finally, it is checked that data

is in whether right side or left side of the hyper plane in order to

classify the new test data [19]. SMOreg uses the sequential

minimal optimization algorithm for training a support vector

classifier. The following equation is used in the optimization

process [20]

max𝛹(𝛼) = ∑ 𝛼𝑖 −1

2∑ ∑ 𝑦𝑖𝑦𝑗𝑘(𝑥𝑖 , 𝑥𝑗)𝛼𝑖𝛼𝑗

𝑁𝑗=1

𝑁𝑖=1

𝑁𝑖=1 (2)

Initial conditions; ∑ 𝑦𝑖𝛼𝑖 = 0, 0 ≤ 𝛼𝑖 ≤ 𝑐 𝑎𝑛𝑑 𝑖 = 1,… , 𝑛𝑁𝑖=1

Where 𝑥𝑖 is trainin sample, 𝑦𝑖 ∈ {−1,+1} is target value, 𝛼𝑖 is

Lagrange multiplier and c is actual cost parameter value. In this

algorithm, the planar kernel function,𝑘(𝑥𝑖 , 𝑥𝑗) = 𝑥𝑖𝑇 ∗ 𝑥𝑗 , was

used.

It preferes to use Gaussian or Polynomial kernels during this

process [21]. In this study, we set the ‘exponent’ property to 2 in

dialog of WEKA software for SMOreg algorithm.This is

necessary, though, to force WEKA to use support vectors. While

SMOreg is using, all the attributes are normalized into binary

attributes [22].

2.3. K-Star Algorithm

K-star is an sample-based classifier that uses entropy based

distance function. Entropy is useful method to evaluate distance

[23]. An instance based algorithm made for symbolic attributes

fails in features of real value thus lacking in incorporated

theoretical base. Therefore this situation makes K-star algorithm

that use entropy based distance function quite important.

Description of K-Star Algorithm [24]:

Let I : infinite set of instances

T is a finite set of transformations on I;

P set of all prefix codes from T* which are determined by σ(the

stop symbol)

Members of T* uniquely define a transformation on

𝐼: 𝑡 ̅(𝑎) = 𝑡𝑛(𝑡𝑛−1)(… . . 𝑡1(𝑎)… . ) Where 𝑡 ̅ = 𝑡1…… . 𝑡𝑛 (3)

A probability function p is defined on T*.It satisfies the following

properties:

0 ≤(�̅�𝑢)𝑝(𝑡)≤ 1 (4)

Σ𝑢 𝑝(𝑡 ̅𝑢) 𝑝(𝑡 ̅) (5)

𝑝(Λ) = 1 (6)

as a consequence, it satisfies the following

Σ𝑝(�̅� ∈ 𝑝�̅�) = 1 (7)

The probability function P* is defined as probability of all paths

from instance a to instance b

P* (b |a) Σ 𝑡 ∈𝑝∶ 𝑡 (𝑎)=𝑏P(𝑡 ) (8)

Σ𝑃 ∗ 𝑏(𝑏|𝑎) = 1 (9)

0 ≤ P*(b|a) = 1 (10)

The K* function is then defined as

K* (b|a) = -log2𝑃∗(𝑏|𝑎) (11)

The basic assumption is that similar examples have similar

classifications. For this reason, it is necessary to define “similar

classifications” and “similar samples” well.

2.4. Multi-Layer Perceptron

The multilayer perceptron (MLP) is a feedback method used the artificial neural network. The Multi-Layer Perceptron consists nodes and neurons [25].

Fig. 2. Multilayer perceptron

Nodes and neurons are placed in three layer (input layer, hidden layer and output layer). The nodes are connected by weights. Each neuron has mathematical function. Each node outputs an activation function applied over the weighted sum of its inputs:

𝑺𝒊 = 𝒇(𝒘𝒊,𝟎 + ∑ 𝒘𝒊,𝒋 × 𝒔𝒋)𝒋∈𝑰 (12)

MLP uses a nonlinear activation function to learn nonlinear function mappings between input and output. These activation functions :

Linear: y = x; (13)

Tanh: y = tanh (x) (14)

Logistic (or sigmoid): y = 1/(1+e-x) (15)

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 314

The input information of a neuron is processed. Subsequently output information of neuron is used as input information for other neuron that is in the next layer. MLP weighs the last relation until the last node is in the last layer. After finding, the method calculates the error and send backward to remodify the model [26]. The neural network used in this study consists of 3 layers.These are the input layer, the hidden layer, and the output layer.The output layer has a single output neuron.When the studies in the literature were examined, it was observed that using 10 neurons in the hidden layer was good in estimating. The number of neurons in the input layer ranged from 1 to 3 according to the groups formed in Table 2 in the 3rd section.

2.5. REP Tree Algorithm

RepTree algorithm prefers to use the regression tree logic. This

algorithm constitutes multiple trees iterations. At the end of the

iterations the best tree is selected among generated trees. The

accuracy of selection is tested by mean square error. Also the

pseudocode of the basic operation of REPTree is shown below in

algorithms 1 and 2 [27].

Data: Dataset D with a set of attributes A

Result: Decision tree using REPTree

if use pruning then

Split D into training data Dt and pruning data Dp;

else

Training data Dt = D;

end

Build a tree using Dt and A as shown in Algorithm 2;

if use pruning then

Reduce error pruning using Dp;

end

Algorithm 1: REPTree algorithm

Data: Dataset D and a set of attributes A

Result: A Decision tree Tree

if no stop condition is reached then

Compute splitting criterion, SC(D,ai), for each attribute ai

∈ A;

Find the best attribute ab according to the splitting

criterion;

Using ab, split D in n subsets;

if max SC > 0 and n > 1 then

foreach of the n subset of Di do

Tree = BuildTree using Di and A;

end

end

else

Tree = Create a leaf using D;

end

Algorithm 2: BuildTree algorithm (REPTree)

RepTree is fast decision tree learning and it builds a decision tree

based on the information gain or reducing the variance [28]. Values

of numeric attributes are classified once by RepTree algorithm.

Reptree algorithm is used to solve both regression and

classification problems.

2.6. M5P Algorithm

M5P is the improved model of M5 algorithm that was created by

Quinlan. The main advantage of M5P algotrithm is that this

algorithm can efficiently handle large number of data sets with

high dimensions [29]. If training set is small, there could be high

classification error rate when comparing with the number of

classes. Parameter setting is not require for M5P algorithm. So this

algorithm does not need knowledge discovery [30]. A M5P model

tree is shown in Figure 3 [31].

Fig. 3. A M5P model tree, ni are split nodes and Mi are the

models

M5P algorithm is fast, simple and have a good accuracy during

process. M5P creates classification and regression trees by using a

multivariate linear regression model. So it can minimize the

variation within a particular sub-space. These model trees

resemble piecewise linear functions. M5P algorithm is also known

as robust algorithm when dealing with missing data.

2.7. Description of Location and Dataset Measurement

The pressure, wind speed, temperature and humidity data that were

used as inputs for both classification and estimation model were

obtained from measurement station that was established.

Measurement station was placed in latitude (N 40o19ʹ58.73ʺ)

longitude (E 36o29ʹ0.28ʺ). Collage photo of measurement station

is shown in Figure 4.

Fig. 4. Application region of measurement station

It is seen that the measurement site is located in a region with

woodland and agricultural areas. The vegetation in the region

consists of short herbaceous plants and rarely straightened trees.

Measurement mast with a height of 12 meters was used in station.

It is shown in Figure 5. Two wind speed sensor and one wind

direction sensor were placed on mast. Pressure, temperature and

humidity sensors were placed in power box. Also data logger was

placed in power box, shown in Figure 6. Solar panel which has 10

W was used to meet energy needs of sensors.

The data (wind speed, pressure, humidity and temperature) were

used in this study were collected at time interval of 10 minute

throughout 2017. The wind speed, pressure, temperature and

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 315

humidity data which are "dat" format were converted into "arff"

format to process in WEKA software. "WEKA" stands for the

Waikato Environment for Knowledge Analysis, is developed by

University of Waikato, New Zealand in 1993. WEKA is a

collection of machine learning algorithms for solving real-world

data mining tasks. It contains tools for data pre-processing,

classification, regression, clustering, association rules and

visualization [32].

Fig. 5. Measurement station

Fig. 6. Power box

3. IMPLEMENTATION OF MACHINE LEARNING ALGORITHMS

At first the missing or inconsistent data in the database were

extracted because this kind of data must be extracted to obtain

successful results. Also it is necessary to remove data repetitions

from the database and increase the data consistency (correctness).

So all data were normalized.

Seven combinations were created from the normalized data

(pressure, humidity and temperature). Groups are given in Table 2.

Table2 Combinations

Combinations

1. Combination Temperature-Humiditiy-Pressure

2. Combination Temperature-Humiditiy

3. Combination Temperature- Pressure

4. Combination Humiditiy-Pressure

5. Combination Temperature

6. Combination Humiditiy

7. Combination Pressure

In order to find the most efficiency method we used 10-fold cross

validation technique which divided data into ten sets of size n/10

(n is number of records). Every data set is tested by using the

remaining sets as its training set. It is shown in Figure 7. The 52560

data were used in the estimation process.

Fig. 7. Schematic representation of 10-fold cross-validation

Fig. 8. Flow chart of estimation model

Anemometers

AIR-X 400W Wind Turbine

Wind

direction

sensors Test Training Training

Training Test Training

Training Training Training

Training Training Training

Training Training ……………………… Training

Training Training Training

Training Training Training

Training Training Training

Training Training Training

Training Training Test

Power supply Pressure sensor Data logger

Temperature sensor Power supply of sensors

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 316

Cross-validation is a technique to evaulate predictive models by

partitioning the original sample into a training set to train the

model, and a test set to evaluate it. In 10 fold cross-validation, the

orginal sample is randomly partitioned into 10 equal size

subsamples.

Of the 10 subsamples, a single subsample is retained as the

validation data for testing the model and the remaining 9

subsamples are used as training data. The cross-validation process

is then repeated 10 times (the folds), with each of the 10

subsamples used exactly once as validation data. The 10 results

from the folds can then be averaged to produce a single estimation.

The advantage of this method is that all observations are used for

both training and validation and each observation is used for

validation exactly once. Our comparing methods that are used to

evaluate algorithms performance are described in detail as follows

chapter 4. The flow chart of estimation model that is established to

examine effects of meteorological parameters is shown in Figure

8. Since the data set we used in this study is very large, only 4464 wind

speed, temperature, humidity and pressure data of January are shown in

Figure 9.

a) b)

c) d)

Fig. 9. a) Wind speed b) Temperature c) Pressure and d) Humidity data of January

4. RESULTS AND DISCUSSION

As explained in Chapter 3, seven combinations were tested for

all algorithms. The performances of the algorithms were

evaluated using Root Mean Square Error (RMSE), Mean

Absolute Error (MAE) and Correlation coefficient (CC)

statistical parameters. Root mean square error is defined as

square root of sum of squares error divided number of

predictions. It is a frequently used to measure the differences

between values predicted by a model and the values actually

observed. It is formulated as given in Equation 2.

𝑅𝑀𝑆𝐸 = √1

𝑁∑ (𝑥𝑖 − 𝑦𝑖)

2𝑁𝑖=1 (2)

𝑥𝑖 is actual wind speed and 𝑦𝑖 is estimated wind speed.

Mean absolute error can be defined as sum of absolute errors

divided by number of predictions. It is measure set of estimated

value to actual value i.e. how close a estimated model to actual

model. MAE is calculated using Equation 3.

𝑀𝐴𝐸 = 1

𝑁 ∑ |𝑥𝑖 − 𝑦𝑖|

𝑁𝑖=1 (3)

Small value of RMSE means better accuracy of model. So,

minimum of RMSE & MAE is better estimation and accuracy.

Correlation coefficient measures the statistical relationship

between the two variables (the actual value and the estimated

value). The correlation coefficient takes values between -1 and

+1. A value close to 1 indicates a good relationship. A value

close to -1 indicates a weak relationship. If the result is 0, it

indicates that there is no relation between the two variables.

Equation of correlation coefficient is given Equation 4.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 317

𝐶𝑜𝑟𝑟𝑒𝑙𝑎𝑡𝑖𝑜𝑛 𝑐𝑜𝑒𝑓𝑓𝑖𝑐𝑖𝑒𝑛𝑡 = ∑ (𝑥𝑖− 𝑥 )(𝑦𝑖− �̅�)𝑁𝑖=1

√∑ (𝑥𝑖− 𝑥 )2𝑁

𝑖=1 √∑ (𝑦𝑖− �̅�)2𝑁

𝑖=1

(4)

There is a direct relationship between the correlation coefficient

and RMSE. If the correlation coefficient is one, the RMSE will

be zero because all points are in the regression line.

The analyzes were carried out in three stages. These are monthly,

annual and seasonal analysis. Firstly, monthly analysis were

performed. Analysis results of first combination are given Table

3 since more lower error values were obtained with estimation

analysis that was made with first combination (temperature-

humiditiy-pressure). Based on the obtained analysis results,

bagging algorithm showed the best performance in all months.

Mean values of MAE, RMSE and CC parameters that were

calculated by using Bagging algorithm are 0.084,0.112 and

0.702 respectively.

Table 3 Statistical analysis results of months

Table 3 Continued

Algorithms

MAE,

RMSE

and CC

Values

Jul Aug Sep Oct Nov Dec

Multilayer

Perceptron

MAE 0.1693 0.1478 0.1178 0.0908 0.1157 0.1040

RMSE 0.2072 0.1834 0.1468 0.112 0.1491 0.1451

CC 0.3509 0.3542 0.3253 0.1292 0.2653 0.2187

SMOreg

MAE 0.1573 0.1416 0.1165 0.0751 0.0992 0.1008

RMSE 0.1880 0.1742 0.145 0.0974 0.1331 0.1445

CC 0.4486 0.3686 0.3464 0.1448 0.4089 0.2505

Kstar

MAE 0.1280 0.1186 0.1032 0.0654 0.0874 0.0851

RMSE 0.157 0.1458 0.128 0.084 0.1136 0.1146

CC 0.676 0.6361 0.5669 0.5289 0.6274 0.6379

Bagging

MAE 0.1044 0.1032 0.0899 0.0588 0.0721 0.069

RMSE 0.1396 0.1332 0.1164 0.0775 0.0983 0.0981

CC 0.7468 0.6991 0.6522 0.6064 0.7289 0.7363

M5P

MAE 0.1207 0.1145 0.0994 0.0642 0.0831 0.0789

RMSE 0.1548 0.1439 0.1253 0.0826 0.1094 0.1081

CC 0.6754 0.6346 0.578 0.5312 0.6467 0.6673

REPTree

MAE 0.1153 0.1104 0.0991 0.0641 0.0797 0.0761

RMSE 0.1557 0.1436 0.1288 0.0848 0.1108 0.1092

CC 0.6779 0.6462 0.5612 0.5214 0.6461 0.6678

It is possible to say that the Bagging algorithm and the first

combination are most efficient (the highest correlation

coefficient, the lowest RMSE and MAE values) in the monthly

estimation. Apart from the Bagging algorithm, also favorable

results were obtained by using the MP5 algorithm in monthly

estimation.

According to months the variations of correlation coefficient

and RMSE values that were calculated by using the first

combination are shown in Fig 10-11. It is seen in Fig. 10-11 that

the highest correlation coefficient value and the lowest RMSE

values were obtained in the bagging algorithm

Fig. 10. The variations of correlation coefficient according to months

Fig. 11. The variations of RMSE values according to months

In the second stage, annual analysis were carried out. Analysis

results according to combinations are shown in Table 4. In the

annual analysis, the lowest RMSE value was obtained in the first

combination (Temperature-Humidity-Pressure). In the three

algorithms (MultilayerPerceptron, SMOreg, Kstar) different

values were obtained in different combinations but the values

that were obtained in the combinations of these algorithms are

worse than the values that were obtained in the first combination

of the bagging algorithm. Good results were obtained again with

the bagging algorithm in all combinations. The worst results

were obtained by multilayer perceptron algorithm.

In the third stage, seasonal analyzes were carried out. The results

of analyzes of all seasons are shown in Table 5-8. In the anaysis

for all seasons, better results were obtained with the Bagging

algorithm and the first combination.

Algorithms

MAE,

RMSE

and CC

Values

Jan

Feb

Mar

Apr

May

Jun

Multilayer

Perceptron

MAE 0.1341 0.1334 0.1458 0.1433 0.1388 0.1390

RMSE 0.1676 0.171 0.1805 0.1808 0.1769 0.1772

CC 0.2374 0.1094 0.216 0.0705 0.1147 0.1612

SMOreg

MAE 0.1116 0.1122 0.1312 0.1333 0.1249 0.1295

RMSE 0.1515 0.1543 0.1705 0.1722 0.1644 0.1699

CC 0.3839 0.2229 0.2225 0.0354 0.1322 0.2225

Kstar MAE 0.0915 0.098 0.1102 0.1111 0.1041 0.1059 RMSE 0.118 0.1255 0.139 0.1417 0.135 0.1372

CC 0.6985 0.6049 0.6053 0.5948 0.5827 0.6237

Bagging MAE 0.0757 0.0811 0.0959 0.0907 0.0855 0.0861

RMSE 0.1011 0.1103 0.126 0.1216 0.1184 0.1185

CC 0.7774 0.6982 0.6802 0.7032 0.6874 0.7215

M5P

MAE 0.0858 0.0929 0.1076 0.1036 0.1001 0.1019

RMSE 0.1122 0.1216 0.1386 0.1337 0.1321 0.1333 CC 0.717 0.6145 0.5909 0.6236 0.5846 0.6279

REPTree

MAE 0.0843 0.0906 0.103 0.1001 0.0983 0.0941

RMSE 0.1146 0.1227 0.1379 0.1371 0.1368 0.1324

CC 0.7087 0.6148 0.6087 0.6111 0.5691 0.6445

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 318

Table 4 Statistical results of annual analysis according to combinations

Algorithms

MAE,

RMSE

and CC

Values

Tem

pera

ture

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Tem

pera

ture

Press

ure

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Press

ure

Multilayer Perceptron

MAE 0.1041 0.1035 0.1041 0.1047 0.1029 0.1035 0.1056

RMSE 0.1327 0.1322 0.1328 0.1335 0.1318 0.1323 0.1348

CC 0.144 0.1321 0.1348 0.1328 0.1317 0.1311 0.0815

SMOreg

MAE 0.0874 0.0892 0.0956 0.0941 0.0941 0.0985 0.0978

RMSE 0.1225 0.1232 0.1226 0.1233 0.1233 0.1234 0.1252

CC 0.2096 0.2704 0.2818 0.2668 0.2668 0.2674 0.2164

Kstar MAE 0.0916 0.0836 0.0922 0.0978 0.0879 0.0898 0.0947

RMSE 0.1153 0.1202 0.1175 0.119 0.1208 0.1212 0.1225

CC 0.4134 0.2917 0.3695 0.3305 0.2764 0.2697 0.2319

Bagging MAE 0.0716 0.0883 0.0767 0.085 0.0914 0.0955 0.0943

RMSE 0.0969 0.1149 0.1026 0.1112 0.1179 0.1205 0.1209

CC 0.6358 0.4099 0.5769 0.4676 0.3567 0.2801 0.2958

M5P

MAE 0.0831 0.0936 0.0869 0.0913 0.0953 0.0959 0.0964

RMSE 0.1076 0.1185 0.1115 0.1159 0.1204 0.1208 0.1216 CC 0.5155 0.3303 0.4598 0.3857 0.283 0.272 0.2491

REPTree

MAE 0.078 0.0916 0.0828 0.0886 0.0941 0.0956 0.096

RMSE 0.1062 0.1186 0.111 0.1157 0.1213 0.1206 0.1228 CC 0.5502 0.3523 0.488 0.4107 0.2971 0.2785 0.2599

Table 5 Statistical analysis results of spring

Algorithms MAE, RMSE

and CC Values

Tem

pera

ture

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Tem

pera

ture

Press

ure

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Press

ure

Multilayer

Perceptron

MAE 0.1424 0.1442 0.1426 0.1435 0.1441 0.1451 0.1424 RMSE 0.1771 0.1786 0.1772 0.1779 0.1786 0.1796 0.1778

CC 0.1367 0.1017 0.1007 0.0978 0.0982 0.095 0.064

SMOreg

MAE 0.124 0.1248 0.124 0.1244 0.1249 0.126 0.1253

RMSE 0.1622 0.1631 0.1621 0.1624 0.1629 0.1637 0.164 CC 0.2319 0.2114 0.2344 0.2269 0.2167 0.1991 0.1804

Kstar

MAE 0.1151 0.1254 0.1206 0.1215 0.1267 0.1272 0.1261

RMSE 0.1446 0.1571 0.1513 0.1521 0.159 0.1587 0.1588 CC 0.4959 0.2897 0.4049 0.3893 0.2475 0.2547 0.2607

Bagging

MAE 0.0926 0.1208 0.101 0.1095 0.1259 0.1252 0.1249

RMSE 0.1238 0.1536 0.1338 0.1429 0.1594 0.1575 0.1597

CC 0.6555 0.3629 0.5771 0.4944 0.2688 0.2768 0.272

M5P

MAE 0.1051 0.1224 0.1126 0.1143 0.1253 0.1255 0.1241

RMSE 0.1356 0.1539 0.1444 0.1459 0.1575 0.1579 0.1567

CC 0.5618 0.3431 0.4733 0.4551 0.2764 0.2681 0.2926

REPTree MAE 0.101 0.1227 0.109 0.1136 0.1261 0.1254 0.1256

RMSE 0.1366 0.1556 0.1449 0.1472 0.1593 0.1577 0.1595

CC 0.5685 0.3346 0.4921 0.4537 0.2522 0.2718 0.2546

Table 6 Statistical analysis results of summer

Algorithms

MAE,

RMSE

and CC

Values

Tem

pera

ture

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Tem

pera

ture

Press

ure

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Press

ure

Multilayer

Perceptron

MAE 0.1441 0.1441 0.1455 0.1442 0.1455 0.1442 0.1492 RMSE 0.1824 0.1823 0.1839 0.1825 0.1838 0.1824 0.1888

CC 0.156 0.1522 0.13 0.1502 0.13 0.1517 0.0068

SMOreg

MAE 0.1342 0.1344 0.1345 0.1342 0.1347 0.1344 0.1382

RMSE 0.1699 0.1702 0.17 0.1701 0.1703 0.1703 0.1775 CC 0.2711 0.2704 0.2667 0.2675 0.266 0.268 -0.0207

Kstar

MAE 0.116 0.1326 0.1217 0.1288 0.1051 0.1363 0.1377

RMSE 0.1469 0.1641 0.154 0.1609 0.1422 0.1675 0.1711 CC 0.5642 0.3353 0.4981 0.3978 0.5755 0.2728 0.2045

Bagging

MAE 0.0797 0.1099 0.0874 0.1051 0.1164 0.1356 0.1267

RMSE 0.1133 0.1457 0.1243 0.1422 0.153 0.1671 0.1644

CC 0.761 0.546 0.6998 0.5755 0.4771 0.2781 0.3415

M5P

MAE 0.1068 0.1283 0.1124 0.1202 0.1326 0.1362 0.135

RMSE 0.1379 0.1604 0.1452 0.1202 0.1646 0.1672 0.1688

CC 0.6099 0.3866 0.5509 0.4719 0.3231 0.2751 0.2417

REPTree MAE 0.0873 0.1184 0.0939 0.113 0.1247 0.1358 0.1312

RMSE 0.1267 0.1574 0.1353 0.1523 0.1625 0.1674 0.1689

CC 0.6906 0.4515 0.6374 0.4979 0.3879 0.2724 0.2917

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 319

Table 7 Statistical analysis results of autumn

Algorithms

MAE, RMSE

and CC

Values

Tem

pera

ture

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Tem

pera

ture

Press

ure

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Press

ure

Multilayer Perceptron

MAE 0.1032 0.1038 0.1037 0.1045 0.1037 0.1044 0.1092

RMSE 0.1269 0.1273 0.1273 0.1281 0.1274 0.128 0.1358

CC 0.1909 0.1742 0.17 0.1641 0.1702 0.1645 0.0203

SMOreg MAE 0.0897 0.0898 0.0898 0.09 0.0898 0.09 0.093 RMSE 0.1135 0.1135 0.1134 0.1141 0.1135 0.1141 0.1201

CC 0.3263 0.3252 0.3269 0.3107 0.326 0.3111 0.1087

Kstar MAE 0.083 0.0893 0.0878 0.0878 0.0901 0.0906 0.0944 RMSE 0.1035 0.1102 0.1087 0.1087 0.1114 0.1119 0.1173

CC 0.4942 0.3679 0.4037 0.4037 0.3452 0.333 0.1456

Bagging

MAE 0.0737 0.0888 0.0848 0.0848 0.0918 0.0902 0.0955

RMSE 0.0956 0.1113 0.1077 0.1077 0.1146 0.1115 0.1199 CC 0.5902 0.3605 0.4273 0.4273 0.2915 0.3379 0.1313

M5P

MAE 0.0795 0.0882 0.086 0.086 0.0895 0.09 0.0943

RMSE 0.1011 0.1096 0.1075 0.1075 0.1109 0.1113 0.1171 CC 0.5212 0.379 0.4195 0.4195 0.3506 0.3413 0.1492

REPTree

MAE 0.0788 0.0878 0.0856 0.0856 0.0898 0.0901 0.0949

RMSE 0.1027 0.1098 0.1083 0.1083 0.1114 0.1113 0.1186

CC 0.5143 0.379 0.4143 0.4143 0.3416 0.342 0.1276

Table 8 Statistical analysis results of winter

Algorithms

MAE,

RMSE

and CC

Values

Tem

pera

ture

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Tem

pera

ture

Press

ure

Hu

mid

ity

Press

ure

Tem

pera

ture

Hu

mid

ity

Press

ure

Multilayer

Perceptron

MAE 0.1211 0.1223 0.1222 0.1222 0.1236 0.1226 0.123

RMSE 0.1551 0.1583 0.1559 0.1562 0.1593 0.1586 0.1568 CC 0.1325 0.0735 0.1277 0.1279 0.0588 0.0679 0.1163

SMOreg

MAE 0.0972 0.0993 0.0974 0.0972 0.0996 0.0995 0.099

RMSE 0.1341 0.1389 0.1342 0.1341 0.1392 0.1393 0.1345

CC 0.3395 0.2367 0.3373 0.3397 0.2278 0.2301 0.3308

Kstar

MAE 0.0894 0.1011 0.0924 0.0966 0.1017 0.1029 0.1002

RMSE 0.1179 0.1334 0.1212 0.1271 0.1345 0.1355 0.1309

CC 0.5594 0.2906 0.5176 0.4352 0.2619 0.2367 0.3617

Bagging MAE 0.0747 0.1004 0.0793 0.0905 0.1036 0.1021 0.1000

RMSE 0.1021 0.1339 0.1073 0.121 0.1371 0.1348 0.1311

CC 0.68 0.2987 0.6372 0.499 0.2329 0.2499 0.3564

M5P MAE 0.084 0.0998 0.0874 0.0929 0.1016 0.1029 0.0989

RMSE 0.1116 0.1321 0.1148 0.1222 0.1338 0.1354 0.1285

CC 0.5974 0.3157 0.5659 0.4788 0.2774 0.2309 0.3846

REPTree

MAE 0.0813 0.1004 0.0852 0.0922 0.1026 0.1021 0.0998

RMSE 0.1122 0.1335 0.1157 0.1231 0.1354 0.1349 0.1303 CC 0.604 0.2984 0.5678 0.4772 0.2459 0.2484 0.3615

a)

b) Fig. 12. The changes of between estimated and measurement values according to algorithms. a) Bagging Algorithm b) Multilayer Perceptron Algorithm

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 320

Summer was chosen to see graphically the variation between the

estimated and measured data. The changes of the 200 measured

and estimated data that were randomly selected for summer are

shown in Fig 12. According to bagging and multilayer

perceptron algorithms, which have the best and lowest

performance, the differences of between Bagging and MLP

algorithms are clearly seen.

Fig. 13. The variations of correlation coefficient according to seasons

Fig. 14. The variations of RMS values according to seasons

According to seasons the variations of the RMSE values and

correlation coefficient that were calculated by using the first

combination are clearly shown in Fig 13-14. In seasonal

analysis, the highest correlation coefficient value and the lowest

RMSE values were all over obtained in the bagging algorithm.

5. CONCLUSION

In this study, Bagging algorithm is proposed to examine the

effect of meteorological parameters on wind speed. Bagging

algorithm has many advantages than the other machine learning

algorithms. Bagging reduces variance or model inconsistency

over diverse data sets from a given distribution, without

increasing bias, which results in a reduced overall generalization

error and enhanced stability.The other benefit of using bagging

is related to the model selection. Since bagging transforms a

group of over fitted neural networks into a better-than perfectly

fitted network, the tedious time consuming model selection is no

longer required. This could even offset the computational

overhead needed in bagging that involves training many neural

networks. Also Bagging is very robust to noise. For that reasons

when the bagging algorithm is compared with other

conventional machine learning algorithms, more efficient results

were obtained with this algorithm in this study. The measured

data were divided into seven combinations to examine the effect

of meteorological parameters and the statistical performances of

the algorithms used with each combination were evaluated

according to RMSE, CC and MAE criteria. The best results were

obtained in combination that were used temperature pressure and

humidity data for wind speed estimation. It was obtained that

temperature and pressure parameters are more effective in wind

speed estimation.

The RMSE value was reduced % 17.81 respect to the

temperature parameter, %19.85 respect to the pressure parameter

and 19.58% respect to the humidity parameter with the

combination used in the annual analyzes. As a result, all

meteorological parameters should be used in correct estimation

of the wind speed. It is aimed to examine the effect of other

meteorological parameters such as altitude, wind direction, air

density in the subsequent studies.

References

[1] REN21, RENEWABLES 2017 GLOBAL STATUS REPORT, 2017,

[Online]. Available: http://www.ren21.net/wp-

content/uploads/2017/06/178399_GSR_2017_Full_Report_0621_

Opt.pdf.; Accessed on June 28 2018.

[2] WEC, Global Wind Statistics, Global Wind Energy Councıl,2016,

[Online].Available:http://www.gwec.net/wpcontent/uploads/vip/G

WEC_PRstats2016_EN_WEB.pdf.; Accessed on June 28 2018.

[3] I. Kırbas, “Short-term multi-step wind speed prediction using

statistical methods and artificial neural networks,” SAKARYA

UNIVERSITY JOURNAL OF SCIENCE, 22 (1), 24-38, 2018.

[4] S. H. Mirhashemi, M. Tabatabayi, “Evaluation ten models of weka

software to predict the potential evapotranspiration month for the

next month in the synoptic weather station Arak,” International

Journal of Agriculture and Crop Sciences, 8(1), 5, 2015.

[5] E. Cadenas, W. Rivera, R. Campos-Amezcua, C. Heard, “Wind

speed prediction using a univariate ARIMA model and a

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 311–321 | 321

multivariate NARX model,” Energies, 9(2), 109, 2016.

[6] M. A. Ansari, N. S. Pal, H. Malik, “Wind speed and power

prediction of prominent wind power potential states in India using

GRNN, In Power Electronics, Intelligent Control and Energy

Systems (ICPEICES),” IEEE International Conference on IEEE ,

1-6, 2016.

[7] M. Zeng, X. Zhang, J. Li, Q. Meng, “Multifractal analysis of short-

term wind speed time series with different sampling frequencies,”

12th World Congress on Intelligent Control and Automation

(WCICA), pp. 3213–3218, 2016.

[8] I. Khandelwal, R. Adhikari, G. Verma, “Time Series Forecasting

Using Hybrid ARIMA and ANN Models Based on DWT

Decomposition,” Procedia Comput. Sci., vol. 48, pp. 173–179,

2015.

[9] M. Khanna, N. K. Srinath, J. K. Mendiratta, “Feature Extraction of

Time Series Data for Wind Speed Power Generation,” 6th

International Conference on Advanced Computing (IACC), pp.

169–173, 2016.

[10] H. G. Lee, M. Piao, Y. H. Shin, “Wind Power Pattern Forecasting

Based on Projected Clustering and Classification Methods,” Etri

Journal, 37(2), 283-294, 2015.

[11]Y. Wanga, H. Mab, D. Wanga, G. Wangc, J. Wua, J. Biand, J.

Liud, “A new method for wind speed forecasting based on copula

theory,” Environmental Research, 160, 365–371, 2018.

[12]R. Velo, P. López, F. Maseda, “Wind speed estimation using

multilayer perceptron,” Energy Conversion and Management, 81,

1–9, 2014.

[13] M. Timur, F. Aydın, T.Ç. Akıncı, “İstanbul Göztepe Bölgesinin

Makine Öğrenmesi Yöntemi ile Rüzgâr Hızının Tahmin

Edilmesi,” Makine Teknolojileri Elektronik Dergisi 8(4), 75-80, (in

Turkish) 2011.

[14]M. Zontul, F. Aydın, G. Doğan, S. Şener, O. Kaynar, “Wind Speed

Forecasting Using REPTree and Bagging Methods in Kirklareli-

Turkey,” J. Theor. Appl. Inf. Technol., 56, 17–29, 2013.

[15] A. Meng, J. Ge, H. Yin, S. Chen, “Wind speed forecasting based on

wavelet packet decomposition and artificial neural networks trained

by crisscross optimization algorithm,” Energy Conversion and

Management,114, 75-88, 2016.

[16] L. Breiman, “Bagging predictors,” Univ. California, Dept. Stat.,

Berkeley, Tech. Rep. 421, 1994.

[17] T. Jiang, J. Li, Y. Zheng, C. Sun, “Improved Bagging Algorithm for

Pattern Recognition in UHF Signals of Partial Discharges,”

Energies, 4, 1087-1101; doi:10.3390/en4071087, 2011.

[18] K. Machová, F. Barčák, P. Bednár, “A Bagging Method using

Decision Trees in the Role of Base Classifiers,” Acta Polytechnica

Hungarica, Vol. 3, No. 2, 2006.

[19] C. Doukas, I. Maglogiannis, P. Tragas, D. Liapis, G. Yovanof,

“Patient fall detection using support vector machines,” International

Federation for Information Processing, vol. 247, pp. 147-156, 2007.

[20] T. W. Kuan, J. F. Wang, J. C. Wang, P. C. Lin, et al., “VLSI Design

of an SVM Learning Core on Sequential Minimal Optimization

Algorithm,” IEEETransactions on Very Large Scale Integration

(Vlsi) Systems, 20(4), 673-683, 2012.

[21] J. Platt, “Fast training of support vector machines using sequential

minimal optimization,” In B. Schölkopf, C. Burges, and A. Smola,

editors, Advances in kernel methods: Support vector learning.

Cambridge, MA: MIT Press. 1998.

[22] A. J. Smola, B. Schölkopf, “A tutorial on support vector regression,”

Statistics and Computing, 14(3), 199-222, 2004.

[23] C. K. Madhusudana, H. Kumar, S. Narendranath, “Condition

monitoring of face milling tool using K-star algorithm and

histogram features of vibration signal,” Engineering Science and

Technology, an International Journal, 19 1543–1551, 2016.

[24] R. Satishkumar, V. Sugumaran, “ Remaining Life Time Prediction

of Bearings Using K-Star Algorithm – A Statistical Approach”,

Journal of Engineering Science and Technology Vol. 12, No. 1

168 – 181, 2017.

[25] A. W. Harley, “An Interactive Node-Link Visualization of

Convolutional Neural Networks,” in ISVC, pages 867-877, 2015

[26] Y. Kumar, G. Sahoo, “Analysis of Parametric & Non Parametric

Classifiers for Classification Technique using WEKA,” I.J.

Information Technology and Computer Science, 7, 43-49, 2012.

[27] D.P.Alonso, S.P. Tejedor, M. Navarro, C. Rad, Á.A. González,

J.F.D.Pastor, “Decision Trees for the prediction of environmental

and agronomic effects of the use of Compost of Sewage Sludge

(CSS),” Sustainable Production and Consumption 12 119 –

133, 2017.

[28]S. Kalmegh, “Analysis of WEKA Data Mining Algorithm

REPTree, Simple Cart and RandomTree for Classification of Indian

News,” International Journal of Innovative Science, Engineering &

Technology, Vol. 2 Issue 2, February, 2015.

[29] A. Behnood, V. Behnood, M. M. Gharehveran, K. E. Alyamac,

“Prediction of the compressive strength of normal and high-

performance concretes using M5P model tree algorithm,”

Construction and Building Materials, 142, 199–207, 2017.

[30] R. Sharma, S. Kumar, R. Maheshwari, “Comparative Analysis of

Classification Techniques in Data Mining Using Different

Datasets,” International Journal of Computer Science and Mobile

Computing, Vol. 4, Issue. 12, December, pg.125 – 134, 2015.

[31] E. K. Onyari, F. M. Ilunga, “Application of MLP Neural Network

and M5P Model Tree in Predicting Streamflow: A Case Study of

Luvuvhu Catchment, South Africa,” International Journal of

Innovation, Management and Technology, Vol. 4, No. 1, February

2013.

[32] R. Meenal, A. I. Selvakumar, “Assessment of SVM, Empirical and

ANN based solar radiation prediction models with most influencing

input parameters,” Renewable Energy, Volume 121, June, Pages

324-343, 2018.

International Journal of

Intelligent Systems and

Applications in Engineering

Advanced Technology and Science

ISSN:2147-67992147-6799 www.atscience.org/IJISAE Original Research Paper

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 322–328 | 322

Lupsix: A Cascade Framework for Lung Parenchyma

Segmentation in Axial CT Images

Hasan Koyuncu*1

Accepted : 31/12/2018 Published: 31/12/2018 DOI: 10.1039/b000000x

Abstract: Lung imaging and computer aided diagnosis (CAD) play a critical role in detection of lung diseases. The most significant part

of a lung based CAD is to fulfil the parenchyma segmentation, since disease information is kept in the parenchyma texture. For this purpose,

parenchyma segmentation should be accurately performed to find the necessary diagnosis to be used in the treatment. Besides, lung

parenchyma segmentation remains as a challenging task in computed tomography (CT) owing to the handicaps oriented with the imaging

and nature of parenchyma. In this paper, a cascade framework involving histogram analysis, morphological operations, mean shift

segmentation (MSS) and region growing (RG) is proposed to perform an accurate segmentation in thorax CT images. In training data, 20

axial CT images are utilized to define the optimum parameter values, and 150 images are considered as test data to objectively evaluate

the performance of system. Five statistical metrics are handled to carry out the performance assessment, and a literature comparison is

realized with the state-of-the-art techniques. As a result, parenchyma tissues are segmented with success rates as 98.07% (sensitivity),

99.72% (specificity), 99.3% (accuracy), 98.59% (Dice similarity coefficient) and 97.23% (Jaccard) on test dataset.

Keywords: Axial Analysis, Computed Tomography, Cascade Framework, Lung Parenchyma, Medical Image Segmentation

1. Introduction

Lung segmentation reserves a significant place before important

objectives like detection of pulmonary nodules, pulmonary

vessels, cancer types and various diseases [1-4]. In medical area,

computed tomography (CT) scans are frequently preferred to

perform a diagnostic about aforementioned subjects. Herein,

thorax CT images constitute some challenging handicaps for an

accurate parenchyma segmentation. These handicaps arise as: 1-)

Similar grey levels between the lung parenchyma and other tissues,

2-) Disease based deformation inside of parenchyma, 3-) Non-

stable shape features of lung tissues. For this purpose, various

segmentation algorithms are asserted in the literature:

Thresholding approaches, Region-based algorithms, Shape-based

methods, Neighboring anatomy-guided techniques and Machine

learning-based structures [5-7].

Thresholding approaches work efficient, once lung intensities

clearly differentiate from the close tissues. However, these

techniques cannot be used on conditions that spatial information

and variability gain importance related to the disease based

deformations. At this point, deformations can sharply change the

grey values in lung [5]. Region-based algorithms (region growing,

fuzzy connectedness, graph cut, etc.) are seen as impressive

techniques containing spatial information with region phenomena.

Region-based methods are actively utilized on homogenous-field

in lung, but the performance of techniques vary on the

segmentation of heterogeneous fields [5]. Shape-based methods

(atlas or model – based methods, active contours, level sets, etc.)

are frequently appealed to segment the abnormal tissues on which

thresholding approaches remain insufficient [5]. Neighboring

anatomy-guided techniques consider neighboring anatomic objects

around lung for separating the lung from objects (organs).

Nevertheless, these techniques cannot be processed on extremely

abnormal or artifacted lung parenchyma [5]. Similarly,

performance of machine learning-based methods can deteriorate

sharply because of the extreme deformation inside of lung

parenchyma [5]. In the literature, aforementioned techniques have

been applied to the lung segmentation, and hybrid techniques are

being designed to handle the parenchyma segmentation with

higher performance [5-8].

Wei et al. [9] developed a system based on improved chain code

and Bresenham algorithm to segment the lungs. Proposed system

performed the task with an accuracy of 95.2%. Dai et al. [10]

designed a pipeline involving graph cuts, Gaussian mixture models

(GMMs) and expectation maximization (EM) to segment the

lungs. According to the results, proposed model obtained the Dice

similarity coefficient (DSC) as 98.74%. Noor et al. [11] realized

lung segmentation with a framework in which thresholding and

morphology based process are used. In experiments, it was

revealed that proposed method performed the segmentation with

high DSC (98.4%). Pulagam et al. [12] utilized the modified

convex hull approach and morphological operators for

segmentation. According to the experiments, proposed approach

obtained 98.62% DSC on overall. Chae et al. [13] generated a

Level-set based approach named MLevel-set that achieved 98.4%

accuracy on lung segmentation. Zhang et al. [14] produced a

hybrid geometric active contour model and attained with 97.7%

DSC on average.

In this paper, a cascade framework is presented to segment the

parenchyma tissue in CT images as more proper than the state-of-

the-art methods. Histogram analysis, morphological operations,

mean shift segmentation (MSS) and region growing (RG) are

handled to design an efficient framework. Proposed system is

named according to the tasks to be realized including the number

of units (six): Lupsix (lung parenchyma segmentation in axial CT).

_______________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________

1 Konya Technical University, Konya, TURKEY

* Corresponding Author: Email: [email protected]

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 322–328 | 323

Performance of Lupsix is evaluated in CT images taken from

LUNA16 dataset, with the use of five different metrics [15-17].

The rest of paper is organized as follows. In Section 2, the utilized

methods and Lupsix are comprehensively explained. In Section 3,

performance evaluation is examined in detail, and visual

perspectives of results are presented to qualify the operation of

system. The interpretations of work are mentioned, and a literature

comparison is tendered. In Section 4, concluding remark is given.

2. Methods

The principle of this study is to obtain the lungs in thorax-axial CT

images in which shape features of lungs varies

uncharacteristically, diseases based deformation can be available

and lung parenchyma can own to close grey levels with the nearby

tissues. This situation requires a robust framework that will not

effected from these handicaps, and the following methods are

utilized to design a robust system:

Gamma correction

Morphological closing

Contrast-limited adaptive histogram equalization

Mean shift segmentation

Region growing

The filling process

2.1. Morphological Algorithms

Gamma correction: Gamma correction (GC) stands for a contrast

enhancement method explicitly revealing the objects in image. GC

is generally utilized to improve the nonlinearity of image and

regarding this, the intensities are rearranged as in (1) [18].

𝑣𝑜𝑢𝑡 = (𝐴. 𝑣𝑖𝑛𝑝𝑢𝑡)𝛾 (1)

vinput, A and γ respectively symbolize the input image, constant and

the power. Mapping process of gamma correction constitutes a

linear transformation in the event that A and γ are equal to “1”, and

pixel intensities are settled within the range [0,1] as in our study

[18].

Morphological closing: Morphological closing (MC) is used to

combine the familiar pixel intensities according to the structuring

element (diamond, disk, line, etc.). For this purpose, closing

process (A•B) is formed as in (2) of which A⊕B and θB

respectively symbolize the dilation of A and B, and erosion of B

[19].

𝐴 • B = (𝐴 ⊕ 𝐵)𝜃𝐵 (2)

As a result of MC, the objects in an image come to the forefront

better than the initial conditions. Hence, MC can be efficiently

preferred on object detection and segmentation tasks [19,20].

Contrast - limited adaptive histogram equalization: Contrast-

limited adaptive histogram equalization (CLAHE) varies from the

adaptive histogram equalization (AHE) with the use of contrast

limiting (CL). Herein, AHE is a transformation of pixels

concerning the histogram and neighboring so as to improve the

irregular contrast distribution. CL represses the increase in noise

level arisen as a result of contrast level increase [21].

Filling process: The filling process is used to fill the areas

connected with each other, and these areas consist of closed shapes

to be filled [22].

2.2. Mean Shift Segmentation

Mean shift segmentation (MSS) defines the 2D features of an

image into the joint spatial-range domain. In operation of MSS,

mode values are considered to rearrange the pixels’ intensities of

different clusters. Two kernels are analyzed to attach the pixels of

an object: 1-) Spatial kernel, 2-) Range kernel. The bandwidths of

kernels play a critical role for the optimum attachment of pixels of

the same cluster (object). MSS algorithm includes three parameters

to approximate the pixels [23]:

Spatial kernel bandwidth (hs)

Range kernel bandwidth (hr)

Convergence criterion threshold (M)

To attach the pixels, MSS operates a multiple-kernel stated in (3)

[23,24].

𝐾ℎ𝑠ℎ𝑟(𝑥) =

𝐶

ℎ𝑠2ℎ𝑟

𝑃 𝑘 (‖𝑥𝑠

ℎ𝑠‖

2

) 𝑘 (‖𝑥𝑟

ℎ𝑟‖

2

) (3)

xs and xr symbolize the vectors concerning with the spatial and

range phenomena; k(x), C and p respectively mean the general

profile, normalization coefficient and dimension. Herein, the

spatial (hs) and range (hr) kernel bandwidths should be carefully

assigned to optimally detect the objects. Besides, the pixels owing

to lower grey values than threshold M, are ignored to perform the

specified-size segmentation. For a detailed explanation about

MSS, please see [23] and [24].

2.3. Region Growing

Region growing (RG) starts with a random or user-defined point

(seed) in image. According to seed intensity, the extension of

pixels continue until similar intensities are not available. For this

purpose, a parameter is needed to decide which pixel value will be

included or excluded. This parameter (tolerance) is regarded to

restrict the extension of pixels. RG algorithm processes step by

step according to the following rules [24]:

Assignment of seed point(s)

The extension is realized which is restricted by tolerance

The extension is performed according to neighbor pixel

intensities until termination is met

2.4. Lupsix Framework

In Lupsix; histogram analysis, morphological operators, MSS and

RG algorithms are chosen to reveal the parenchyma tissue, and to

design a robust system against aforementioned handicaps. For this

purpose, Lupsix runs the operation steps in Figure 1:

1. GC differentiates the regions inside of thorax CT image.

At this point, lung parenchyma differs from the nearby

tissues like ribs, subcutaneous fats, etc.

2. MC is performed with disk type structuring element,

since this type element remains as the most proper one

in terms of fitting to the concavity of lungs. After MC is

processed, the parenchyma intensities come to closer

grey levels for an accurate object segmentation.

3. CLAHE sharpens the outer edges of parenchyma.

However, this situation puts forward more sensitive

changes inside of parenchyma, and some parts of lung

(like nodules, vessels, etc.) are sharpened too.

4. MSS attaches the tissue parts of parenchyma to reveal

and to smooth the parenchyma more explicit. Regarding

this, thorax CT image becomes more convenient before

object segmentation.

5. RG is fulfilled using two user-defined seed points, and

lung parenchyma is segmented with the use of MSS

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 322–328 | 324

output (image containing smoothed and attached

objects).

6. The filling process is realized to combine the small gaps

comprising of circumscribed areas within the

parenchyma tissues.

7. System output is obtained by multiplying the output of

filling process (binary mask) with the output of GC.

Fig. 1. Flowchart of proposed framework

In Lupsix, five parameters should be examined in detail to perform

an accurate segmentation:

Tolerance in RG

Size of structuring element in MC

Spatial kernel bandwidth (hs)

Range kernel bandwidth (hr)

Threshold in MSS (M)

3. Results and Discussions

In this section; parameter adjustment, training & test results, visual

results, interpretations and literature comparison are presented in a

comprehensive and brief manner. All experiments were performed

in Matlab software on a personal computer with 2.80 GHz CPU,

16 GB RAM and GeForce GTX 1050Ti graphic card. Performance

of framework is evaluated on 170 CT images taken from LUNA16

database [15-17]. Five statistical metrics (sensitivity, specificity,

accuracy, DSC, Jaccard) are considered to evaluate the

performance of system:

Sensitivity = TP/ (TP+FN) (4)

Specificity = TN/ (TN+FP) (5)

Accuracy = (TP+TN)/(TP+TN+FP+FN) (6)

DSC = (2*TP) / ((2*TP)+FP+FN) (7)

Jaccard = TP / (FN+FP+TP) (8)

These metrics are preferred according to their efficiency for

evaluation of segmentation performance [24, 25]. TP, TN, FP, and

FN respectively specify the true positive, true negative, false

positive and false negative results among trials [20,24,25].

3.1. Parameter Settings of Lupsix

There exist three rules for design of a robust framework: 1)

Formation of units step by step according to the needs, 2)

Examination of parameters in detail by using the train data, 3)

Detection of the optimum values by using the satisfactory results

specified with a tolerance [25]. To actualize this, parameter values

stated in Table 1 are handled for an in-depth analysis.

Table 1. Parameter adjustments of Lupsix

Parameter Examination Values and Ranges

Tolerance 10, 20, 30, 40, 50, 60

Size of structuring element 2, 3, 4, 5, 6

[hs,hr]

[8,8], [16,16], [24,24], [32,32], [40,40],

[48,48], [16,8], [16,24], [16,32], [16,40],

[16,48], [24,8], [24,16], [24,32], [24,40], [24,48], [32,8], [32,16], [32,24], [32,40],

[32,48], [40,8], [40,16], [40,24], [40,32],

[40,48], [48,8], [48,16], [48,24], [48,32], [48,40]

M 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7,

0.8, 0.9, 1, 1.5, 2

The bandwidths of spatial (hs) and feature (hr) kernels are

examined as dual and fold of “4” [13, 24]. Performance of Lupsix

sharply deteriorates once tolerance, size of structuring element and

M are selected bigger than 60, 6 and 2, respectively. Also different

values and conditions stated in Table 1, have been handled, but the

results of these trials were unsuitable to present. In brief, Table 1

presents the reliable adjustments having meaningful results on

parenchyma segmentation.

20 thorax-axial CT images are considered to adjust the parameters

of framework, and Table 2 is formed in which red results

symbolizes the best ones. In Table 2, blue results symbolize the

second best results having to a deviation about -0.8% (Jaccard)

from the best results. The second best results are utilised to

generalize the operation of framework and to fix the parameter

values. Because, parameters cannot be fixed in case only the best

results are considered.

Table 3 presents the repetition number of all parameter choices,

and optimum operation conditions are revealed as the result of 20

different trials. According to Table 3;

Tolerance value of “50” was efficient on all trials, and

can be regarded as the optimum tolerance.

Once size of structuring element is assigned as “4”,

optimum results were achieved by means of five

statistical metrics on 18/20 trials. Concerning this,

optimum size of element can be assigned as “4”.

The optimum duality of [hs,hr] is [40,32] at which the

best results were achieved on all trials.

The Lupsix output ensued the optimum results when M

= 0.2 or M = 0.3 on all trials. However, the optimum

threshold was “0.3” since the best results of this choice

were more than the choice of “0.2”.

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 322–328 | 325

3.2. Experimental Results and Analysis

For an objective and challenging comparison, we have used 20 CT

images on training, while data number is chosen as 150 on test.

Besides, we control whether every operator of system is necessary

or not. Table 4 shows the results with / without operators of Lupsix

on a randomly chosen CT image.

According to Table 4, every operator upgrades the performance of

system, and the best performance cannot be achieved without these

operators as seen in the gaps between with / without results. RG

algorithm play a selector role on the choice of lung regions and

concerning this, it’s not included as a section in Table 4.

Table 2. Optimum parameter values obtained on training dataset

Image Tolerance of RG Size of Structuring

Element [hs,hr] Threshold in MSS

Train_001 30,40,50,20 3,2,4,5 [48,24],[32,32],[40,40],[24,32],[24,40], [24,48],[32,24],[32,40],[32,48],[40,24],

[40,32],[40,48],[48,16],[48,32],[48,40]

1,1.5,2,0.1,0.2,0.3,0.4,

0.5,0.6,0.7,0.8,0.9

Train_002 40,30,50,60 4,3,5,6

[48,16],[24,24],[32,32],[40,40],

[32,16],[32,24],[40,16],[40,24], [40,32],[48,24],[48,32],[48,40]

0.3,0.1,0.2,0.4,0.5,

0.6,0.7,0.8,0.9,1

Train_003 50,30,40,60 4,2,3,5,6

[32,40],[32,32],[40,40],[16,40],

[16,48],[24,40],[24,48],[32,48], [40,32],[48,24],[48,32],[48,40]

0.4,0.5,0.6,0.1,0.2,0.3,

0.7,0.8,0.9,1,1.5,2

Train_004 50,30,40 3,2,4

[32,40],[24,24],[32,32],[40,40],[24,16],[24,32],

[24,40],[24,48],[32,16],[32,24],[32,48],[40,16], [40,24],[40,32],[48,16],[48,24],[48,32],[48,40]

0.1,0.3,0.4,0.2,0.5,0.6,

0.7,0.8,0.9,1,1.5,2

Train_005 60,30,40,50,70 2,3,4,5,6 [24,40],[16,16],[24,24],[32,32],[40,40],[16,24],[16,32],[16,40],

[16,48],[24,8],[24,16],[24,32],[24,48],[32,8],[32,16],[32,24],[32,40],

[40,8],[40,16],[40,24],[40,32],[48,8],[48,16],[48,24],[48,32],[48,40]

0.1,0.2,0.3,0.4,0.5,0.6,

0.7,0.8,0.9,1,1.5,2

Train_006 40,30,50 2,3,4,5 [40,40],[32,32],[16,48],[40,32] 0.8,0.9,1,0.2,0.3,0.4,

0.5,0.6,0.7,1.5

Train_007 60, 50 4, 3 [32,32],[40,40],[32,24],[40,24],[40,32] 0.3,0.1,0.2,0.4,0.5,

0.6,0.7,0.8,0.9

Train_008 50,30,40,60,70 3,2,4

[48,40],[32,32],[40,40],[48,48],[16,32],

[16,40],[16,48],[24,32],[24,40],[24,48], [32,40],[32,48],[40,32],[40,48],[48,32]

0.1,0.2,0.3,0.4,0.5,

0.6,0.7,0.8,0.9,1

Train_009 40,20,30,50 6,2,3,4,5

[40,32],[8,8],[16,16],[24,24],[32,32],[40,40],[16,24],[16,32],[16,40],

[16,48],[24,8],[24,16],[24,32],[24,40],[24,48],[32,8],[32,16],

[32,24],[32,40],[40,8],[40,16],[40,24],[48,8],[48,16],[48,24],[48,32]

2,0.1,0.2,0.3,0.4,0.5, 0.6,0.7,0.8,0.9,1,1.5

Train_010 70,20,30,40,50,60 3,2,4,5,6 [32,48],[40,40],[48,48],[24,48],[32,40],[40,32],[40,48],[48,40] 0.5,0.6,0.7,0.1,0.2,0.3,

0.4,0.8,0.9,1,1.5,2

Train_011 70,30,40,50,60 2 [48,40],[40,40],[32,40],[32,48],[40,24],[40,32],[48,24],[48,32] 0.4,0.1,0.2,0.3,0.5,0.6,

0.7,0.8,0.9,1,1.5

Train_012 40,20,30,50 2,3,4,5 [48,16],[16,16],[24,24],[32,32],[16,8],[16,24],[16,32], [24,8],[24,16],[24,32],[32,8],[32,16],[32,24],[40,8],

[40,16],[40,24],[40,32],[48,8],[48,24],[48,32]

0.1,0.2,0.3,0.4,0.5,0.6,

0.7,0.8,0.9,1,1.5,2

Train_013 70,20,30,40,50,60 4,2,3,5

[48,48],[24,24],[32,32],[40,40],[24,32],[24,40],

[32,16],[32,24],[32,40],[40,16],[40,24],[40,32], [48,16],[48,24],[48,32],[48,40]

0.3,0.2

Train_014 40,20,30,50,60 3,2,4,5 [48,32],[16,16],[24,24],[32,32],[40,40],[16,24],[16,32],[16,40], [16,48],[24,16],[24,32],[24,40],[24,48],[32,16],[32,24],[32,40],

[32,48],[40,16],[40,24],[40,32],[48,16],[48,24],[48,40]

0.1,0.2,0.3,0.4,0.5,0.6,

0.7,0.8,0.9,1,1.5,2

Train_015 20,30,40,50,60,70 3,2,4 [40,40],[32,32],[24,32],[32,40],[40,32],[40,48],[48,32],[48,40] 0.3,0.1,0.2,0.4,0.5,0.6,

0.7,0.8,0.9,1,1.5

Train_016 70,20,30,40,50,60 2

[40,48],[24,24],[32,32],[40,40],[48,48],[16,32],[16,40],[16,48],

[24,32],[24,40],[24,48],[32,24],[32,40],[32,48],

[40,16],[40,24],[40,32],[48,24],[48,32],[48,40]

0.2,0.1,0.3,0.4,0.5,0.6, 0.7,0.8,0.9,1,1.5,2

Train_017 40,30,50 2,3,4 [32,24],[32,32],[24,32],[32,40],[40,24],[40,32],[48,24],[48,32] 0.1,0.2,0.3,0.4,0.5, 0.6,0.7,0.8,0.9,1

Train_018 30,40,50 4,2,3,5,6 [40,40],[16,40],[16,48],[24,48],[32,40],[32,48],

[40,24],[40,32],[48,24],[48,32],[48,40] 0.3,0.1,0.2,0.4,0.5,0.6,

0.7,0.8,0.9,1,1.5,2

Train_019 50, 40 4, 5 [24,32],[32,32],[40,40],[24,40],[32,24],[32,40],[40,32] 0.3,0.1,0.2,0.4,0.5,0.6,

0.7,0.8,0.9,1,1.5

Train_020 60,30,40,50 3,2,4,5,6 [32,32],[24,24],[40,40],[24,16],[24,32],[24,40],

[24,48],[32,16],[32,24],[32,40],[32,48],[40,16],[40,24],[40,32]

,[40,48],[48,16],[48,24],[48,32],[48,40]

0.4,0.5,0.6,0.7,0.8,0.9,

1,1.5,0.1,0.2,0.3,2

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 322–328 | 326

Training and test results are presented in Table 5. According to

Table 5, Lupsix achieves to better sensitivity, specificity, DSC and

Jaccard rates on test dataset meaning that parameter settings of

framework were obtained as optimal, and Lupsix can perform the

task of segmentation on new images (test dataset) with more

reliable performance.

Table 4. Performance evaluation of Lupsix operators

Performance

Without Sections

(With / Without)

Accuracy DSC Jaccard

GC 99.63 / 16.13 98.85 / 27.77 97.73 / 16.13

MC 99.63 / 99.55 98.85 / 98.59 97.73 / 97.21 CLAHE 99.63 / 37.20 98.85 / 33.93 97.73 / 20.43

MSS 99.63 / 16.13 98.85 / 27.77 97.73 / 16.13

The Filling Process 99.63 / 99.51 98.85 / 98.46 97.73 / 96.97

Table 5. Performance evaluation of Lupsix on datasets

Dataset Sensitivity Specificity Accuracy DSC Jaccard

Train 97,91 99,67 99,33 98,17 96,41

Test 98,07 99,72 99,30 98,59 97,23

3.3. Visual Results and Interpretations

Figure 2 shows the visual results of Lupsix for different conditions.

As seen in Figure 2;

Lupsix operators can effectively divide the parenchyma

tissue from other tissues in thorax CT image,

Lupsix can work on different shaped and sized

parenchyma tissues by achieving high segmentation

performance,

Lupsix segments different parenchyma tissues having

various grey levels. At this point, proposed framework

can perform the segmentation on both homogenous and

heterogeneous tissues,

Lupsix is not effected from the deformations stated in

lung parenchyma,

As a result, an efficient framework is designed to be

utilized before disease determination and classification,

tumor segmentation and specification, nodule detection,

vessel segmentation and abnormality categorization.

Table 6 presents the comparison of Lupsix with the state-of-the-art

techniques on lung segmentation. According to the results, it’s seen

that DSC rate of Lupsix remain promising among other studies in

the literature, and the best accuracy is achieved by proposed

system. Consequently, a robust segmentation system is presented

to the literature for segmentation of lung parenchyma.

Table 6. The literature comparison on lung segmentation

Study Techniques Accuracy Dice

Wei et al.

[9]

System based on improved chain

code and Bresenham algorithm 95.2% -----

Dai et al.

[10]

Pipeline involving graph cuts,

Gaussian mixture models

(GMMs) and expectation

maximization (EM)

----- 98.74%

Noor et

al. [11]

Framework including thresholding and morphology

based segmentation coupled with

feedback

----- 98.4%

Pulagam

et al. [12]

The modified convex hull

approach and morphological

operators

----- 98.62%

Chae et

al. [13] MLevel-set 98.4% -----

Zhang et

al. [14]

Hybrid geometric active contour

model ----- 97.7%

This study Lupsix 99.3% 98.59%

Table 3. Repetition number of parameter values

Parameter

Name

Parameter

Value

Repetition Number among All

Trials (20 Training Images)

Tolerance

10 0

20 8

30 18

40 19

50 20

60 12

70 7

Size of

Structuring

Element

2 17

3 17

4 18

5 13

6 7

[hs,hr]

[8,8] 1

[16,16] 4

[24,24] 9

[32,32] 17

[40,40] 18

[48,48] 4

[16,8] 1

[16,24] 4

[16,32] 6

[16,40] 7

[16,48] 8

[24,8] 3

[24,16] 6

[24,32] 13

[24,40] 11

[24,48] 11

[32,8] 3

[32,16] 8

[32,24] 13

[32,40] 16

[32,48] 10

[40,8] 3

[40,16] 9

[40,24] 14

[40,32] 20

[40,48] 6

[48,8] 3

[48,16] 9

[48,24] 14

[48,32] 16

[48,40] 14

Threshold

0.1 18

0.2 20

0.3 20

0.4 19

0.5 19

0.6 19

0.7 19

0.8 19

0.9 19

1 18

1,5 15

2 11

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 322–328 | 327

4. Conclusions

In this paper, lung parenchyma segmentation is handled, and an

efficient & semi-automated framework is designed to obtain the

parenchyma in thorax-axial CT images.

For this purpose, Lupsix is formed by using histogram analysis,

morphological approaches, MSS and RG algorithms to efficiently

segment the parenchyma. As a result, a robust framework is

achieved challenging to the state-of-the-art techniques in the

literature.

According to the experiments, it’s seen that Lupsix segmented the

lung parenchyma with higher success rates on test data than

training dataset. Also we kept the training data (20 images) very

low according to the test data (150 images) so as to perform an

objective and challenging comparison. On the other hand, a

detailed examination is performed to find the optimum operation

conditions of Lupsix. Concerning this, high segmentation

performance is achieved by the combinatory work of necessary

operators in Lupsix.

In future work, Lupsix will be utilized as the segmentation part of

a classification framework in which disease classification will be

realized using the parenchyma tissue obtained as the output of

Lupsix.

References

[1] S. Zhou, Y. Cheng, and S. Tamura, “Automated lung segmentation and

smoothing techniques for inclusion of juxtapleural nodules and

pulmonary vessels on chest CT images,” Biomed. Signal Process., vol.

13, pp. 62-70, 2014.

[2] S. Shen, A. A. Bui, J. Cong, and W. Hsu, “An automated lung

segmentation approach using bidirectional chain codes to improve

nodule detection accuracy,” Comput. Biol. Med., vol. 57, pp. 139-149,

2015.

[3] M. Obert, M. Kampschulte, R. Limburg, S. Barańczuk, and G. A.

Krombach, “Quantitative computed tomography applied to interstitial

lung diseases,” Eur. J. Radiol., 2018.

[4] M. Z. Ur Rehman, M. Javaid, S. I. A. Shah, S. O. Gilani, M. Jamil, and

S. I. Butt, “An appraisal of nodules detection techniques for lung

cancer in CT images,” Biomed. Signal Process., vol. 41, pp. 140-151,

2018.

[5] A. Mansoor, U. Bagci, B. Foster, Z. Xu, G. Z. Papadakis, L. R. Folio,

J. K. Udupa, and D. J. Mollura, “Segmentation and image analysis of

abnormal lungs at CT: current approaches, challenges, and future

trends,” RadioGraphics, vol 35, no. 4, pp. 1056-1076, 2015.

[6] B. A. Skourt, A. El Hassani, and A. Majda, “Lung CT Image

Segmentation Using Deep Neural Networks,” Procedia Comput. Sci.,

vol. 127, pp. 109-113, 2018.

Fig. 2. Outputs of Lupsix operators

International Journal of Intelligent Systems and Applications in Engineering IJISAE, 2018, 6(4), 322–328 | 328

[7] X. Liao, J. Zhao, C. Jiao, L. Lei, Y. Qiang, and Q. Cui, “A

segmentation method for lung parenchyma image sequences based on

superpixels and a self-generating neural forest,” Plos One, vol. 11, no.

8, e0160556, 2016.

[8] T. Doel, D. J. Gavaghan, and V. Grau, “Review of automatic

pulmonary lobe segmentation methods from CT,” Comput. Med. Imag.

Grap., vol. 40, pp. 13-29, 2015.

[9] Y. Wei, G. Shen, and J. J. Li, “A fully automatic method for lung

parenchyma segmentation and repairing,” J. Digit. Imaging, vol. 26,

no.3, pp. 483-495, 2013.

[10] S. Dai, K. Lu, J. Dong, Y. Zhang, and Y. Chen, “A novel approach of

lung segmentation on chest CT images using graph cuts,”

Neurocomputing, vol. 168, pp. 799-807, 2015.

[11] N. M. Noor, J. C. Than, O. M. Rijal, R. M. Kassim, A. Yunus, A. A.

Zeki, M. Anzidei, L. Saba, and J. S. Suri, “Automatic lung

segmentation using control feedback system: morphology and texture

paradigm,” J. Med. Syst., vol. 39, no.3, 22, 2015.

[12] A. R. Pulagam, G. B. Kande, V. K. R. Ede, and R. B. Inampudi,

“Automated lung segmentation from HRCT scans with diffuse

parenchymal lung diseases,” J. Digit. Imaging, vol. 29, no.4, pp. 507-

519, 2016.

[13] S. H. Chae, H. M. Moon, Y. Chung, J. Shin, and S. B. Pan, “Automatic

lung segmentation for large-scale medical image management,”

Multimed. Tools Appl., vol. 75, no.23, pp. 15347-15363, 2016.

[14] W. Zhang, X. Wang, P. Zhang, and J. Chen, “Global optimal hybrid

geometric active contour for automated lung segmentation on CT

images,” Comput. Biol. Med., vol. 91, pp. 168-180, 2017.

[15] S. Armato III, G. McLennan, L. Bidaut, M. McNitt-Gray, C. Meyer,

A. Reeves, and L. Clarke, “Data from LIDC-IDRI,” The cancer

imaging archive, 2015

[16] S. G. Armato et al., “The lung image database consortium (LIDC) and

image database resource initiative (IDRI): a completed reference

database of lung nodules on CT scans,” Med. Phys., vol. 38, no.2, pp.

915-931, 2011.

[17] K. Clark et al., “The Cancer Imaging Archive (TCIA): maintaining

and operating a public information repository,” J. Digit. Imaging, vol.

26, no. 6, pp. 1045-1057, 2013.

[18] P. Charles, “Digital Video and HDTV Algorithms and Interfaces,”

Morgan Kaufmann Publishers, San Francisco, 2003.

[19] R. C. Gonzalez, R. E. Woods, S. L. Eddins, “Digital image processing

using MATLAB,” McGraw Hill Education, New York, 2009.

[20] H. Koyuncu, R. Ceylan, M. Sivri, H. Erdogan, “An Efficient Pipeline

for Abdomen Segmentation in CT Images,” J. Digit. Imaging, vol. 31,

no.2, pp. 262-274, 2018.

[21] K. Zuiderveld, “Contrast limited adaptive histogram equalization,” in:

Graphics gems IV, P. S. Heckbert, Ed. San Diego, CA, USA:

Academic Press Professional Inc., 1994, pp. 474–485.

[22] P. Soille, “Morphological Image Analysis: Principles and

Applications,” Springer-Verlag, 1999, pp. 173-174.

[23] D. Comaniciu, P. Meer, “Mean shift: A robust approach toward feature

space analysis,” IEEE T. Pattern Anal., vol. 24, no. 5, pp. 603-619,

2002.

[24] H. Koyuncu, R. Ceylan, H. Erdogan, M. Sivri, “A novel pipeline for

adrenal tumour segmentation,” Comput. Meth. Prog. Bio., vol. 159, pp.

77-86, 2018.

[25] S. W. Franklin, S. E. Rajan, “Computerized screening of diabetic

retinopathy employing blood vessel segmentation in retinal images,”

Biocybern. Biomed. Eng., vol. 34, pp. 117-124, 2014.