Consumer Profile and Tipping Habits. A Romanian Framework Using Fuzzy Method
by Ada Mirela Tomescu, Olimpia I. Ban
Abstract: Tipping is an
interesting consumer behavior, because the tips are voluntary payments given
after services are completed. Our paper tries to establish the Romanian sample
profile and their perception regarding tipping phenomena. The Romanian consumer
is difficult to investigate because of resistance to being fully honest in this
problem. Thus the present study aim is to identify whether there are differences
between respondents` attitude toward idea of tipping in various units, and
whether these attitudes are influenced by respondent incomes, this last
objective it was accomplished using Fuzzy method. A thesis could be: a higher
income leads to a more positive attitude towards the idea of giving tips;
another is to record the profile and the attitude to this issue of a certain
sample of Romanian consumers.
Keywords:
Fuzzy method, Interaction, Romanian consumer, Tips
Full Paper, pp. 1-8
Parametric Modeling for Advanced Architecture
by M. Stavric, O. Marina
Abstract: Computer
simulation of evolutionary processes is already well established technique for
the study of environmental, biological and economic dynamics. Use of algorithms
for generation of virtual entities that will develop its functional and formal
properties within the non-linear process of adaptation of complex system is a
foundation for new point of view in understanding architecture and urban
environment. This paper describes parametric approach in architectural design
through elaboration of shift in paradigms in architecture that has brought to
the idea of use of parametric modeling with emphasis on two different groups of
parametric software and presents the possibilities of generative algorithms in
modeling architectural form and development of cities and urban design.
Keywords:
Geometry, parametric design, generative algorithms urbanism, architectural
design
Full Paper, pp. 9-16
by Tan Shiang-Yen, Long Yoon Foo, Rosnah Idrus
Abstract: The purpose of
this paper is to assess the public's perception and acceptance toward the
proposed mobile tagging system, which can be used by consumers to retrieve
product information of Genetically Modified Food (GMF) in the market. In the
proposed mobile tagging system, consumers can use the camera on their mobile
phone to capture the image of the Quick Response (QR) codes on the product
packaging and send to the server for decoding. Subsequently, the server sends
back the details of the GMF product in the form of Multimedia Message Service
(MMS). Findings of this study indicate that albert not much the public is
currently using mobile tagging service, but they possess positive attitude to
find out more and to try the mobile tagging services. Most importantly,
acceptance of the proposed mobile tagging service for retrieving product
information of genetically modified food is considerably high. Additionally, the
finding suggests that marketing campaign is vital to increase awareness among
the consumers, and customer education is important to improve the acceptance
level toward the mobile tagging service. This study is one of the pioneer
studies in mobile tagging system. It provides the feasibility analysis of
applying mobile tagging system in genetically modified food area and assesses
the marketability of the system.
Keywords:
2D Codes, Genetically Modified Food, Mobile Tagging, Quick Response (QR) Code
Full Paper, pp. 17-25
Statistical Causality and Stochastic Dynamic Systems
by Ljiljana Petrovic
Abstract: In this paper we
consider a problem (that follows directly from realization problem): how to find
a possible states (even minimal) of a stochastic dynamic system S1 with known
outputs, provided it is in a certain causality relationship with another
stochastic dynamic system S2 whose states (or some information about them) are
given. This paper is continuation of the papers (Gill and Petrovi´c 1987),
(Petrovi´c 1996) and (Petrovi´c 2005).
Keywords:
Hilbert space, causality, stochastic dynamic system, realization.
Full Paper, pp. 153-156
The HOSVD Based Domain and the Related Image Processing Techniques
by Andras Rovid, Laszlo Szeidl, Peter Varlaki
Abstract: In the framework
of the paper an improved method for image resolution enhancement is introduced.
The enhancement is performed through another representation domain, where the
image function is expressed with the help of polylinear functions on higher
order singular value decomposition (HOSVD) basis. The paper gives a detailed
description on how to determine the polylinear functions corresponding to an
image and how to process them in order to obtain a higher resolution image.
Furthermore, the proposed approach and the Fourier-based approach will be
compared from the point of view of their effectiveness.
Keywords:
Image resolution, HOSVD, approximation, numerical reconstruction
Full Paper, pp. 157-164
Achieving the Efficiency for Information Systems - Possibilities and Prospects
by D. Litan, M. Velicanu, L. Copcea Teohari, M. Teohari, A. M. Mocanu Virgolici, I. Surugiu, O. Raduta
Abstract: Nowadays companies
cannot cope with competition successfully unless they have performing
information systems. Therefore, the need for efficient information systems is
very high. The efficiency can be achieved by using various methods/ techniques
and technologies “to build” the information system. In this paper we are going
to present and analyze to what extent the new technologies and
methods/techniques: methods of integrating the information systems; techniques
of databases optimization; “special” types of databases (mobile, multimedia and
spatial databases); data warehouses, have increased or could increase the
efficiency of information systems.
Keywords:
Database optimization, data warehouse, Enterprise Application Integration,
mobile databases, multimedia databases, spatial databases, Oracle Warehouse
Builder
Full Paper, pp. 165-179
Nonlocal Symmetries for a Family Benjamin-Bona-Mahony-Burgers Equations. Some Exact Solutions
by M. S. Bruzon, M. L. Gandarias
Abstract: In this work the
nonlocal symmetries of a family Benjamin-Bona-Mahony-Burgers equations are
studied. The partial differential equation written as a conservation law can be
transformed into an equivalent system by introducing a suitable potential. The
nonlocal symmetry group generators of original partial differential equation can
be obtained through their equivalent system. We have proved that the
nonclassical method applied to this system leads to new symmetries, which are
not solutions arising from potential symmetries of the Benjamin-
Bona-Mahony-Burgers equations. We also have derived traveling wave solutions for
the Benjamin-Bona-Mahony-Burgers equations by using a direct method. Among them
we nd a solution which describes a kink solution.
Keywords:
Nonlocal symmetries, Partial differential equations, Nonlinear equations, MAXIMA
Software, kink solutions
Full Paper, pp. 180-187
Capacity Recovery of Very Noisy Optical Quantum Channels
by Laszlo Gyongyosi, Sandor Imre
Abstract: The number of
efficient approximation algorithms for quantum informational distances is very
small, because of the special properties of quantum informational generator
functions and of asymmetric quantum informational distances. If we wish to
analyze the properties of quantum channels using today’s classical computer
architectures, an extremely efficient algorithm is needed. The capacity recovery
of very noisy communication channels cannot be imagined for classical systems,
and this effect has no analogue in classical systems. The capacity recovery of
very noisy quantum channels makes it possible to use two very noisy
optical-fiber based quantum channels with a positive joint capacity at the
output. Here we define a fundamentally new approach of capacity recovery of very
noisy, practically completely useless optical quantum channels. We show an
algorithmic solution to the capacity recovery problem, and provide an efficient
algorithmic solution for finding the set of recoverable very noisy optical
quantum channels. The calculations are based on the asymptotic classical
capacity of the quantum channel.
Keywords:
Capacity Recovery, Noisy Optical Communications, Quantum Channels, Quantum
Communications
Full Paper, pp. 188-199
Software Quality Evaluation: User's View
by Anas Bassam AL-Badareen, Mohd Hasan Selamat, Jamilah Din, Marzanah A. Jabar, Sherzod Turaev
Abstract: Nowadays, software
products rapidly increased and it is usage not limited to specific people or
corporation. It is now used in the most of human life activities. Therefore, the
quality of the software product is increasingly being important and the users
demanding higher quality than ever before. However, most of this research is
focused on the internal/development view of quality. Hence, in software
development, strong attention must be given to the user’s satisfaction.
Recently, the studies intend to understand the users’ perspective of the
software quality. In this study, we intend to discuss the characteristics of the
software products that influence the users’ satisfaction on software quality.
Based on the well-known software quality models and the emotion of the software
users, a model of software quality evaluation based on users’ views is proposed.
Keywords:
Software Quality, Quality Model, Software Evaluation, User Perspective
Full Paper, pp. 200-207
Effectiveness of Stemming and n-Grams String Similarity Matching on Malay Documents
by Tengku Mohd T. Sembok, Zainab Abu Bakar
Abstract: There are two main
classes of conflation algorithms, namely, string-similarity algorithms and
stemming algorithms. String-similarity matching algorithms, bi-grams and
tri-grams, are used in the experiments conducted on Malay texts. Malay stemming
algorithms used in the experiments is developed by Fatimah et al. Inherent
characteristics of n-grams on Malay documents are discussed in this paper.
Retrieval effectiveness experiments using several variations of combinations
between n-grams and stemming algorithms are performed in order to find the best
combination. The variations experimented are: both nonstemmed queries and
documents; stemmed queries and nonstemmed documents; and both stemmed queries
and documents. Further experiments are then carried out by removing the most
frequently occurring n-grams. Besides using dice coefficients to rank documents,
inverse document frequency (idf) weights are also used. Interpolation technique
and standard recall-precision functions are used to calculate recall-precision
values. It is found that using combined search, n-gram matching and stemming,
improves retrieval effectiveness. Removing the most frequently occurring n-gram
that appears in about 46% of the words also improve the retrieval effectiveness.
Keywords:
Information retrieval, string similarity matching, stemming algorithms
Full Paper, pp. 208-215
Effective Algorithms for Improving the Performance of Search Engine Results
by G. Poonkuzhali, R. Kishore Kumar, R. Kripa Keshav, K.Thiagarajan, K.Sarukesi
Abstract: Search engine has
become an important tool in todayΤs world for searching various data like text,
audio, video and images. While searching for information many users end up with
irrelevant information causing a waste in user time and accessing time of the
search engine. So to narrow down this problem, many researchers are involved in
web mining, still the algorithms developed by them contains nearly 30% of
outlaid content such as irrelevant and redundant information. In this paper, two
Statistical approaches based on Proportions (Z-test hypothesis) and chi square
test (T-test) is developed for mining this outlaid content. Also comparative
studies between these two methods are presented. Elimination of this outlaid
content during a searching process improves the quality of search engines
further. The results show that the system easily provides relevancies and
delivers dominant text extraction, supporting users in their query to
efficiently examine and make the most of available web data sources.
Experimental results revealed that statistical approach produces better results
than chi square test.
Keywords:
Chi Square Test, critical value, degrees of confidence, relevant, test
statistic, web document
Full Paper, pp. 216-223
Sharing Behaviors in Games and Social Media
by Harri Ketamo
Abstract: User generated
media highlights sharing: Sharing videos, images and texts in social media, as
well as sharing character outfits and maps in games. However, behavior is one of
the aspects that are not shared. The aim of this study is to show how user
generated behaviors can be shared in different types of games and social media
applications. The examples about sport games and educational games are based on
MindFarm AI technology that enables end users to construct human-like behavior
by only teaching. MindFarm emulates the human way to learn: According to
cognitive psychology of learning, our thinking is based on conceptual
representations of our experiences and complex relations between these concepts
and experiences. Phenomena when the mental structure change is called learning.
In the near future user-generated behaviors can be developed and shared, as all
other user-generated content. According to examples on this study, game- and
media developers can design extensions that enable users to easily construct
behaviors. From a game consumer point of view, the most interesting part is in
developing behaviors, sharing them and finally playing with them, or against
them.
Keywords:
Artificial Intelligence, Games, Learning, Social Media, User Generated Content
Full Paper, pp. 224-232
Deformations of Monic Polynomial Matrices. Analysis of Perturbation of Eigenvalues
by M. Isabel Garcia-Planas, Sonia Tarragona
Abstract: In this work we
study behavior of a multiple eigenvalue of the monic polynomial family P(λ) as
well as we study of behavior of a simple eigenvalue of a family of 1-degree
singular
polynomial matrices representing families of singular linear systems.
Keywords:
Polynomial matrix, Eigenvalues, Perturbation
Full Paper, pp. 233-239
Monocriteria and Multicriteria based Placement of Reactive Power Sources in Distribution Systems
by Whester J. Araujo, Petr Ya. Ekel, Rafael P. Falcao Filho, Illya Kokshenev, Henrique S. Schuffner
Abstract: This paper
presents results of research on formulating and solving the problem of capacitor
placement in distribution systems within the framework of monocriteria and
multicriteia models. The application of the multicriteria approach is aimed,
first of all, at overcoming the difficulties of simultaneous observation of
contradictory constraints for the upper and lower permissible voltage limits at
different buses of distribution systems as well as other important conditions in
operating capacitors. The solution of the capacitor placement problem in the
monocriteria statement is based on applying the generalized algorithms of
discrete optimization. The solution of the problem in the multicriteria
statement is based on combining the Bellman-Zadeh approach to decision making in
a fuzzy environment with the application of the generalized algorithms of
discrete optimization. The paper results are illustrated by computational
experiments with a real distribution system.
Keywords:
Discrete optimization, Distribution systems, Muticriteria decision making,
Reactive power compensation
Full Paper, pp. 240-248
Integrated Study of Automated Checkout Shopping Trolley
by Koh Peik See, Tan Shiang-Yen, Nasuha Lee Abdullah, Rosnah Idrus
Abstract: Recently,
Automated checkout shopping trolley technologies have taken an advance step than
the previous introduced self-checkout system in terms of greater convenience,
ease-of-use and greater efficiency. As this technology is considerable new in
the market of developing countries, hence it is vital to explore the perception
of consumers toward the use of automated checkout trolley. The purpose of this
paper is to examine the perceptions and expectations of both enterprise users
and end users toward the proposed checkout system for hypermarket in the context
of Malaysia, as a typical paradigm of developing countries. Findings from this
study indicate that both consumers and retailers possess positive attitude
toward the automated checkout trolley system, and are enthusiastic to try this
emerging technology. The main concerns of the users are the secureness of
payment and the variability of payment methods, which forms their core criteria
in determining whether to adopt the system. This study provides an understanding
about the practicality and market condition of automated trolley system in the
developing countries.
Keywords:
Technology-based self-service (TBSS), Automated checkout, Market survey,
Technology and innovation management
Full Paper, pp. 249-257
by I. Marin Carrion, E. Arias Antunez, M. M. Artigao Castillo, J. J. Miralles Canals
Abstract: Nonlinear time
series analysis is a powerful methodology that permits to predict the temporal
evolution of some kinds of dynamical systems from characteristic quantities,
such as the minimal embedding dimension or the maximal Lyapunov exponent. In
fact, one of the most important goals of nonlinear analysis of experimental time
series is the prediction. The subspace identification methods provides a good
framework to model a system, both deterministic and stochastic, in a easily way.
In order to make predictions, we propose a method which combine the minimal
embedding dimension obtained by the method of false nearest neighbors and a
model estimated by means of a subspace identification method. The results, in
terms of predicted error, show the reliability of this new approach.
Keywords:
Nonlinear Time Series Analysis, False Nearest Neighbors Method, Subspace
Identification Methods, Prediction
Full Paper, pp. 258-265
The Quickest Maximum Dynamic Flow of Minimum Cost
by Mircea Parpalea, Eleonor Ciurea
Abstract: This article
states and solves the multi-criteria maximum flow problem in discrete dynamic
networks for the case of two objective functions. This represents a
generalisation of the maximum flow of minimum cost problem for the case of
minimizing the travelling cost (minimum cost flow) and travelling time (quickest
flow). The approach is actually based on generating efficient extreme points in
the objective space by iteratively solving a series of maximum flow problems
with different single objective functions. Each time, the dynamic flow is
augmented along a cheapest (minimum cost) path from the source node to the sink
node in the time-space network avoiding the explicit time expansion of the
network.
Keywords:
Bi-criteria flow, Discrete dynamic network, Dynamic maximum flow, Successive
shortest path
Full Paper, pp. 266-274
Time Varying Causality between Stock Exchanges in the CEE Region
by Cristiana Tudor
Abstract: This paper
analyzes the evolution of linkages and causality between six Central and Eastern
European (Bulgaria, Czech Republic, Hungary, Poland, Romania, Russia) and the
USA stock exchanges. The effects of the recent global financial crisis and
possible changes in inter-market relationships as a result of the crisis are
topics of special interest in this research. For this purpose, the panel data
sample (daily observations for the 2006-2009 time period) is divided into two
sub-periods corresponding to the pre-crisis and crisis period. In order to
separate the two sub-periods, a secondary investigation is conducted which shows
that the beginning of July 2007 is the moment when the global financial crisis
began to show its full manifestations on international stock exchanges. The
study concludes that stock markets in the CEE region have became increasingly
integrated during crisis, while before the crisis the markets appear to be
segmented, as both contemporaneous correlations and causality relationships are
mostly insignificant. Also, before the crisis CEE markets were significantly
influenced by innovations in the USA market, thus explaining why they were
affected heavily by the crisis, which has managed to spread immediately in the
region. As far as the risk-adjusted performance is concerned, the Czech market
realized the best risk-adjusted performance due to its average rate of return
and low risk, s followed by the Romanian, Russian and Polish stock markets,
while the lowest risk-adjusted performance, as represented by the coefficient of
variation was found in the case of the US stock market.
Keywords:
Granger causality, contemporaneous correlation, risk-adjusted performance,
Central and Eastern European Stock exchanges
Full Paper, pp. 27-34
by D. Litan, A. Mocanu, S. Olaru, A. Apostu
Abstract: Market research...
a real treasure which can trigger performance in any business, but most of the
times it seems unexploited at its true value. The marketing analysis brings
information about competitors, clients, business partners and so on, guiding the
business towards success. Because during the last years the information
technologies have been applied in so many aspects of our lives, they have also
been used for marketing research. The most used marketing research method is the
questionnaire, that is why many of the well known companies from this field have
left the “real” environment and they are focusing now to the “online”
environment. Therefore, today one can not do a marketing analysis without
specialized IT tools (for the online surveys or for data analysis). In the
current paper, we search and offer some answers regarding the information
technologies used, or which could be used, for developing and implementing
information systems for marketing research. In the same time, on the one hand,
we are opening new perspectives and propose new ways of considering the
marketing research: used to design an e-government system about the citizens
needs, expectations or preferences; and on the other hand: online survey systems
“seen” like a first step for a large scale usage of evoting concept.
Keywords:
E-government, geographical information system, information security, information
technology, market research
Full Paper, pp. 35-45
The Image Recognition of Brain-Stem Ultrasound Images with using a Neural Network based on PCA
by Jiri Blahuta, Tomas Soukup, Petr Cermak
Abstract: This paper shows
how to solve the recognition of ultrasound brain-stem images. Our work is based
on PCA method that is very useful and known method for image processing. We will
demonstrate a solving with C application, Gnumeric spreadsheet as a standard
solution. Also we will use an artificial neural networks (ANN) for this problem
and we will compare a results. The ANN are generally very usable for image
processing. It has been demonstrated with NeuroSolutions software that is very
sophisticated simulator of ANN with PCA multilayer (ML) NN topology.
Keywords:
PCA, ultrasound, neural network, image processing
Full Paper, pp. 46-54
by Pavel Jirava, Jiri Krupka, Miloslava Kasparova , Jan Mandys
Abstract: This article is
focused on the problem of the data acquisition and analysis of data about
quality of life of hearing disabled people in the region. Data were obtained
from disabled citizens of Pardubice region, Czech Republic. To obtain the data
was used a questionnaire survey. For processing the obtained data were used the
statistical methods, selected visualization techniques and data mining method.
The achievements of this research are relations and links influencing quality of
life of disabled people and basic characteristics of hearing disabled people.
This is the basis for future social policy making in the Pardubice region. The
methods used in solving the problem can be implemented also in other regions (in
the level Nomenclature of Units for Territorial Statistics 3).
Keywords:
Quality of life, people with disabilities, data mining, cluster analysis,
regional development, sustainable development, questionnaire survey.
Full Paper, pp. 55-62
Analysis of Complexity of One-dimensional Chaotic Maps with Entropy Characteristic
by Sen Pei, Zhiming Zheng
Abstract: Chaotic phenomena
are widespread in a variety of research fields. In this paper, we investigate
the complexity of one-dimensional chaotic maps with entropy characteristic.
First we analyze the dynamics of one-dimensional chaotic maps. Then we analogize
the iteration of maps to a one-sided shift transformation, and give a
probability explanation of Lyapunov exponent by Ergodic Theory. We use the
entropy of corresponding shift to reflect the complexity of these maps,
providing a new way to characterize complexity of chaotic systems.
Keywords:
Chaos, Complexity, Entropy, Ergodicity, Lyapunov Exponent
Full Paper, pp. 63-70
AES Algorithm Adapted on GPU Using CUDA for Small Data and Large Data Volume Encryption
by Tomoiaga Radu Daniel, Stratulat Mircea
Abstract: The goal of this
paper is to study the possibility of using alternative computing solution in
cryptography, the use of a graphic processing unit in non graphical
calculations. We tried to use the graphic processing unit as a cryptographic
coprocessor in order to obtain more computing power and better runtime for AES.
In this paper we present an implementation of AES on NVIDIA GPU using CUDA. The
results of our tests show that the CUDA implementation can offer speedups of
almost 40 times in comparison to the CPU. The tests are conducted in two
directions: running the tests on small amount of data that is located in memory
and a big amount of data that are stored in files on hard drives and as news,
the access time for the information on the hard disk is added to the encrypting
time.
Keywords:
Benchmark, AES, CUDA, GPU, Cryptography
Full Paper, pp. 71-81
Derivations and Iterated Skew Polynomial Rings
by Michael Gr. Voskoglou
Abstract: Two are the
objectives of the present paper. First we study properties of a differentially
simple commutative ring R with respect to a set D of derivations of R. Among the
others we investigate the relation between the D-simplicity of R and that of the
local ring RP with respect to a prime ideal P of R and we prove a criterion
about the D- simplicity of R in case where R is a 1-dimensional (Krull
dimension) finitely generated algebra over a field of characteristic zero and D
is a singleton set. The above criterion was quoted without proof in an earlier
paper of the author. Second we construct a special class of iterated skew
polynomial rings defined with respect to finite sets of derivations of a ring R
(not necessarily commutative) commuting to each other. The important thing in
this class is that, if R is a commutative ring, then its differential simplicity
is the necessary and sufficient condition for the simplicity of the
corresponding skew polynomial ring.
Keywords:
Derivations, Differentially simple rings, Finitely-generated algebras, Iterated
skew polynomial rings, Simple rings
Full Paper, pp. 82-90
by J. Alejandro Betancur
Abstract: Nowadays, in
automobile industry are found many working situations, in which the use of an
articulated arm that supports the activities developed by operators is
inevitable; out side the assembly line of a car, activities such as searching
for parts needed to produce a vehicle, handling the cars tires, among others,
are performed manually in many different companies. In this work is exposed the
design of a robot that helps the operator to develop those activities, which is
determined by a mechanical arm, made up by four (4) segments and four (4)
joints, located on a mobile platform with two (2) wheels, each controlled by an
independent engine. Through this work is intended to cover mathematics and
physics present in the above mentioned device, taking advantage of the
computational power symbolic of the software MAPLE. Although this development is
focused on the automotive sector, applications of this kind of device in other
fields of industry are innumerable.
Keywords:
Arm, Automobile, Denavit, Electromechanical, Euler, Hartenberg, Langrange
Full Paper, pp. 91-100
Considerations Regarding Formal Languages Generation Using Labelled Stratified Graphs
by Tudor (Preda) Irina-Valentina
Abstract: The concept of
labelled stratified graph was introduced in order to obtain the concept of
knowledge base with output and various applications of this concept were
presented ever since. This paper studies another application of this structure:
generating formal languages by means of labelled stratified graphs. Various
mechanisms to define and generate formal languages are known and we show that we
can obtain different types of languages such as: regular languages or context
sensitive language. We also give an example of context sensitive language, but
not a context-free language (according to Chomsky hierarchy), that can be
generated by labelled stratified graphs. The concepts introduced in this paper
can initiate a possible research line concerning the generative power of the
formal languages generated by labelled stratified graphs.
Keywords:
Context-free language, interpretation, labelled graph, labelled stratified
graph, Peano algebra, regular language
Full Paper, pp. 101-108
DiaSys – A Dialogue System based on Conditional Knowledge in Client-Server Technology
by Cristina Zamfir (Tudorache)
Abstract: A conditional
schema is a graph-based structure which is able to represent conditional
knowledge. In [1] the structure of a conditional schema was introduced and in
[2] the inference mechanism corresponding to the conditional schema
representations was developed. The Knowledge Representation and Reasoning (KR&R)
is based on the idea that propositional content can be rigorously represented in
formal languages, such that the resulted representations can be productively
reasoned over by humans and machines [16]. Conditional Knowledge Representation
and Reasoning represents a new brand of KR&R, for which several formalisms have
been developed. The DiaSys application based on conditional knowledge was
presented in [4]. In this paper we describe the evolution of that dialogue
system into a client-server application. The conditional schema will be defined
and exemplified. The server and the client will be presented step by step in two
separate sections. The two sections are oriented specifically towards the task
that the described part of the application performs.
Keywords:
Client-server, Dialog System, Knowledge base, Natural language, XML
Full Paper, pp. 109-116
MAGS – An Approach Using Multi-Objective Evolutionary Algorithms for Grid Task Scheduling
by Miguel Camelo, Yezid Donoso, Harold Castro
Abstract: Grid task
scheduling problem has been a research focus in grid computing for the past
years. Some Deterministic, Heuristics or Metaheuristic scheduling approaches
have been proposed to solve this NP-complete problem. However, these algorithms
do not take the Multi-Objective nature of Grid Computing performance into
account. In this paper we present a Multi-Objective approach using Evolutionary
Algorithm (MOEA) to efficiently solve such kind of scheduling problems. Our
proposal is based on NSGA-II MOEA algorithm combined with a set of Heuristics in
different evolutionary operators which allow a fast convergence to optimal (or
near-optimal) solutions. The results obtained by our proposed algorithm were
compared and evaluated against Mono-Objective and Multi-Objective algorithms
used for Grid task scheduling. The main contributions of this paper are the
proposed mathematical model, the optimization model and the algorithm to solve
it. Additionally we show the effectiveness and robustness of the proposed
algorithm.
Keywords:
Evolutionary Algorithms, Grid, Task Scheduling, Heuristics, Metaheuristics,
Meta-Scheduler, Multi-Objective Optimization, NP-Complete, NSGA-II, Pareto Front
Full Paper, pp. 117-126
Scaled Image Edge Detection Based on The Total Variation Functional
by Peter Ndajah, Hisakazu Kikuchi
Abstract: We present total
variation anisotropic edge detection. We derive the total variation functional
from measure theory, distribution theory and vector gradient method. The
Euler-Lagrange equation of the total variation functional gives a steady state
equation. The steady state equation acts as an anisotropic filter on an image.
The total variation filtered images are compared to Laplacian filtered images. A
subsequent application of the zero crossing algorithm works quite well for the
traditional Marr-Hildreth method but gives poorer results for the total
variation filtered images. It was found that thresholding methods work better
and saves computational time. Also, our results show that total variation edge
detection overcomes some drawbacks associated with the Marr-Hildreth method.
Keywords:
Anisotropic, Euler-Lagrange, LoG, Total Variation
Full Paper, pp. 127-136
Technologies for Development of the Information Systems: from ERP to e-Government
by D. Litan, A. Apostu, L. Copcea (Teohari), M. Teohari
Abstract: Although some ERP
(Enterprise Resource Planning) software products are available on the market,
they do not entirely fit each and every type of business. That is the main
reason why companies prefer to develop their own ERP type of systems, thus
eliminating “barriers” such as long time and high costs required to develop the
software, compared with the shorter time and lower costs of just purchasing the
ERP system and customizing it to the targeted business. However, in the short
run, lower cost will eventually result into losses in the long run, due to that,
one the one hand, the system is not perfectly adapted to business specifics and,
on the other hand, customizing an ERP package can influence implementation of
the standard benefits of an integrated system. Therefore, many times, the best
choice for a company would be to develop an ERP type of system that caters
perfectly for the needs of the organization. In this paper, we will proceed to
the analysis, presentation and comparison of the current most important
information methodologies (for business analysis and for IT projects –
people-oriented methodologies) and databases/ data warehouses which are being
used in development of the ERP systems. Also, another aspect that we will
consider, in our paper is to examine how some of the technologies used in the
development of ERP systems can be applied in the development of e-government
type systems.
Keywords:
Data warehouse, e-government, Enterprise Resource Planning, information
methodology, query optimization, information system, information technology
Full Paper, pp. 137-152
Efficient ID-Based Multi-Proxy Multi-Signature Scheme Based on CDHP
by Rajeev Anand Sahu, Sahadeo Padhye
Abstract: Proxy signature is
useful in situation when a user wants to authorize an agent called proxy signer
to sign any document on his behalf. Multi-proxy multi-signature is one of the
primitives of proxy signature. Bilinear pairing makes the system efficient and
provides an ease in computation. In this paper, we propose an ID-based
multi-proxy multi-signature scheme from bilinear pairings based on Computational
Diffie-Hellman Problem (CDHP), replacing the certificate generation of Li and
Chen’s scheme by delegation generation. Our scheme is proxy protected and
computationally more efficient than the ID-based multi-proxy multi-signature
scheme of Li and Chen. We also analyze the security properties of our scheme and
show that the proposed scheme satisfies all the security requirements of a safe
proxy signature scheme.
Keywords:
Bilinear pairings, Cryptography, ID-based signature scheme, Multi-proxy
multi-signature, Security analysis
Full Paper, pp. 275-282
Elman Networks for the Prediction of Inventory Levels and Capacity Utilization
by F. Harjes, B. Scholz-Reiter, A. Kaviani Mehr
Abstract: Today's production
processes face an increase in dynamics and complexity. Therefore, production
control techniques face a demand for continuous advancement. Methods from the
field of artificial intelligence, such as neural networks, have proven their
applicability in this area. They are applied for optimization, prediction,
classification, control and many other production related areas. This paper
introduces an approach using Elman Networks for the workstation-specific
prediction of inventory levels and capacity utilization within a shop floor
environment. It includes the selection of the appropriate network architecture,
the determination of suitable input variables as well as the training and
validation process. The evaluation of the proposed approach takes place by means
of a generic shop floor model.
Keywords:
Artificial neural networks, Elman networks, predictive control, shop floor
production
Full Paper, pp. 283-290
by Zamalia Mahmud
Abstract: Postgraduate
students who are involved in the analyiss of survey research data are expected
to have a sound knowledge in statistics. However, statistics learners should
firstly inculcate interest, feel the importance of statistics and instill
positive attitude toward learning statistics. Based on Rasch Unidimensional
Measurement Model, this study shall investigate the attitude, importance and
knowledge in statistics among postgraduate students. The investigation involved
a small classroom experiment measuring students’ perceived attitude prior to and
after attending the statistics remedial classes and then their perceived
knowledge on various statistical tools and concepts. The validity of the
response patterns were measured and whether their responses can be taken as
measuring the latent trait or otherwise. This paper also describes how Rasch
probabilistic model is used to measure the latent trait of students’ attitude in
response to the agent or items measuring the trait. This involves a calibration
between person ability to respond towards the five-point Likert scale items and
whether the items function in unison to form a single underlying pattern. Logit
scores as derived from the Rasch measurement model, Person-Item Distribution map
and other related Rasch statistics are executed to endorse the trait. The study
shows that student demonstrate a higher cognitive positive attitude toward
learning statistics after attending the remedial classes with most post-attitude
items superseding the pre-attitude items during the calibration. However, there
are traces of misfit responses as identified by the fit indices due to unusual
or/and inappropriate responses which are highlighted in this paper.
Keywords:
Attitude towards learning statistics, Pre-post attitude Rasch probabilistic
model, PIDM, Person-Item misfit, Perceived Importance and Knowledge
Full Paper, pp. 291-298
by Florentin Serban, Maria Viorica Stefanescu, Silvia Dedu
Abstract: This paper present
a combination of classification theory with risk estimation theory to determine
the best assets through those that we analyzed. We present also building to the
efficient frontier for a portfolio with 2 or 3 assets We use a data analysis
method to obtain two classes of assets and then we estimate the risk of each
asset corresponding to each class. Thus, we get the best two assets among those
considered risky for which we build the efficient frontier if we consider that
the portfolio consists of these two risky assets and a risk-free asset. To
illustrate the effectiveness of the method used, we present a case study that
refers to the Bucharest Stock Exchange. We construct the efficient frontier,
based on the correlation of the best stocks that we have obtained through data
analysis (for classification), and by assessing the loss distribution (for risk
assessment), taking into account that the portfolio contains an asset without
risk.
Keywords:
Efficient frontier, Optimization, Portfolio selection, Principal component
analysis, Risk estimation
Full Paper, pp. 299-306