Journal of Computers, Vol 5, No 12 (2010), 1775-1778, Dec 2010
doi:10.4304/jcp.5.12.1775-1778

Guest Editorial

Syed Mahfuzul Aziz, Vijayan K. Asari, M. Alamgir Hossain, Mohammad Ataul Karim, Mariofanna Milanova

Abstract


Despite being subjected to an economic downturn that is yet to recover, computers and information technology (IT) continue to drive and influence innovation, growth and development in almost all sectors of the world economy. The gains in efficiency and productivity obtained from the optimum use of emerging computer and IT technologies are having lasting impacts on the global society. These technologies are also playing a vital role in human health care and in enhancing our knowledge and understanding of the environment by supporting advanced modelling underpinned by high performance algorithms and processing paradigms. At the core of the continuing uptake and impact of computer and IT technologies is the research that is conducted worldwide in these fields. During the past decade the IEEE International Conference on Computer and Information Technology (ICCIT) has been successful in bringing together academics, researchers, IT professionals and managers to disseminate and discuss the up to date research findings in these fields.

This Special Issue presents selected papers from the twelfth conference of the series (ICCIT 2009) held during December 21-23, 2009 at the Independent University Bangladesh. The first one was held in Dhaka, Bangladesh, in 1998. Since then the conference has grown to one of the largest computer and IT related research conferences in the South Asian region, with participation of academics and researchers from many countries around the world. Authors are required to submit full length papers to the conference for review. A double blind review process is followed whereby each paper is reviewed by at least two independent reviewers of high international standing. The acceptance rate of papers in recent years has been around 32% or less. This is an indication of the quality of work the papers accepted for the conference demonstrates. Starting in 2008 the proceedings of ICCIT are included in IEEExplore. The 2009 ICCIT papers are no exception.

In 2009, a total of 473 full papers were submitted to the conference of which 150 were accepted after reviews conducted by an international program committee comprising 84 members from 16 countries with assistance from 96 reviewers. This was tantamount to an acceptance rate of 31.7%. Form the 150 papers accepted for the conference only 14 highly ranked papers were invited for this Special Issue. The authors were invited to enhance their conference papers significantly, with at least 30% extension, and submit the same for review. Of those only eight papers survived the review process and have been selected for inclusion in this Special Issue. The authors of these papers represent academic and/or research institutions from Bangladesh, Canada, Japan, Norway and the United States. These eight papers cover six domains of computing namely, application-driven algorithms, data mining, neural networks, distributed systems, content sharing and high-performance digital processors.

Fourteen reviewers from five countries have assisted the guest editors in reviewing the papers submitted to the Special Issue during two rounds of review process. They have given immensely to the process by responding to the guest editors in the shortest possible time and by dedicating their valuable time to ensure that the Special Issue contains high-quality papers with significant contributions. The guest editors would like to express their sincere gratitude to all the reviewers, namely: M. Shafiul Alam, Abdul S. Awwal, Abul K. M. Azad, Abusaleh M. Jabir, Joarder Kamruzzaman, Mozammel H. A. Khan, A. S. Madhukumar, Babu Mailachalam, Kazi Muheymin, Hau Ngo, Mukaddim K. Pathan, Vinod A. Prasad, Praveen Sankaran and M. Hasan Shaheed.

The first paper by S. Akter and M.H.A. Khan uses quantum-inspired evolutionary algorithm (QEA) for multiple-case outlier detection in multiple linear regression models. In this study, a Bayesian information criterion based fitness function incorporating extra penalty for a number of potential outliers has been used to identify the most appropriate set of potential outliers. The QEA is shown to overcome the effects of smearing and masking, and effectively detects the most appropriate set of outliers. The next paper by M.A. Rahman and M.M. Akbar presents the problems associated with existing cluster based hierarchical solutions to cope with the large number of participants in distributed applications that often use peer-to-peer computing. These solutions exploit the idea of a coordinator/leader of cluster, and therefore the fault tolerance is adversely affected when a coordinator fails. To address this problem the authors propose a cluster based network architecture of two layers of hierarchy and a hierarchical permission based algorithm, which is free from the use of coordinator.

The first of two papers covering data mining by K.S.N. Ripon, A. Rahman and G.M.A. Rahaman considers eliminating duplicate records especially when the records are domain-dependent so as to minimize over-representation of data. The investigators propose a novel domain-independent technique for better reconciling the similar-duplicate records. They introduce ideas for making similar-duplicate detection algorithms faster and more efficient. Finally they propose an algorithm that incorporates these techniques for similar-duplicate detection into a domain-independent environment. In the second paper, R. Paul and A.S.M.L. Hoque consider a search efficient representation of healthcare data based on the Health Level Seven (HL7) standard. They have proposed a search efficient data model, the optimized entity attribute value (OEAV), for physical representation of medical data. They have used OEAV to model the observation class of reference information model (RIM), and used relational model for the remaining RIM classes. They have shown that OEAV is more search-efficient and occupy less storage space compared to conventional EAV.

In the fifth paper, M.A.H. Akhand, P.C. Shill and K. Murase present a data sampling based neural network ensemble method where the individual networks are trained on the union of the original training set and a set of some artificially generated examples. After each network is trained, the method checks whether the trained network is suitable to ensemble or not, and absorbs the network based on the ensemble performance associated with it.

The sixth paper of this Special Issue, by R.I. Rafat and K. Sakib, explores consistency models for distributed shared memory (DSM). This study considers a new consistency model in DSM named Last Update Consistency (LUC) model, where the model uses a logical clock counter to keep the DSM consistent. In this model, multiple nodes perform READ operations over the same block at a time. For WRITE operation over the same block, only the last modification exists and the earlier WRITE operations are treated as obsolete and are discarded. The authors show that the proposed model effectively reduces unnecessary network traffic.

M.M.R. Chowdhury, S. Alam, Z. Iqbal and J. Noll then in the seventh paper explore the issue of security in content sharing. Social relations are used extensively as access constraints to secure the shared content, however, only relations cannot provide personalized and granular enough access control. To mitigate the problems, the authors propose an access authorization model incorporating diverse real life social relations and associated attributes such as trust, distance of relations and frequency of interactions, and demonstrate practical applications of such a model.

The last paper of this special issue pertains to the design of high-performance digital processors. R. Tajallipour, M.A. Islam and K.A. Wahid present an efficient algorithm to compute base-10 logarithm of a decimal number. The algorithm uses 64-bit floating-point arithmetic, and is based on a digit-by-digit iterative computation that does not require look-up tables, curve fitting, decimal-binary conversion or division operations. The architecture is pipelined and implemented on to the Xilinx Virtex2p FPGA. According to the authors it is the first FPGA prototype of its kind that uses a 64-bit (decimal 16-digit) precision. 



References



Full Text: PDF


Journal of Computers (JCP, ISSN 1796-203X)

Copyright @ 2006-2014 by ACADEMY PUBLISHER – All rights reserved.