Sunday, 26 August 2018 03:52

Google N-Gram Viewer does not Include Arabic Corpus! Towards N-Gram Viewer for Arabic Corpus

Izzat Alsmadi1 and Mohammad Zarour2

1Department of Computing and Cyber Security, Texas A&M University, USA

2Information Systems Department, Prince Sultan University, KSA

Abstract: Google N-gram viewer is one of those newly published Google services. Google archived or digitized a large number of books in different languages. Google populated the corpora from over 5 million books published up to 2008. This Google service allows users to enter queries of words. The tool then charts time-based data that show the frequency of usage of query words. Although Arabic is one of the top spoken language in the world, Arabic language is not included as one of the corpora indexed by the Google n-gram viewer. This research work discusses the development of large Arabic corpus and indexing it using N-grams to be included in Google N-gram viewer. A showcase is presented to build a dataset to initiate the process of digitizing the Arabic content and prepare it to be incorporated in Google N-gram viewer. One of the major goals of including Arabic content in Google N-gram is to enrich Arabic public content, which has been very limited in comparison with the number of people who speak Arabic. We believe that adopting Arabic language by Google N-gram viewer can significantly benefit researchers in different fields related to Arabic language and social sciences.

Keywords: Arabic language processing, corpus, google N-gram viewer.

Received May 7, 2015; accepted September 20, 2015

 


Sunday, 26 August 2018 03:50

Capacity Enhancement Based on Dynamically

Adapted PF Scheduling Algorithm for LTE Downlink System

Mohammed Abd-Elnaby, Mohamad Elhadad, and El-Sayed El-Rabaie

Department of Electronics and Electrical Communications, Menoufia University, Egypt

Abstract: Orthogonal Frequency Division Multiplexing (OFDM) with dynamic scheduling and resource allocation is a key component of most emerging broadband wireless access networks such as Worldwide Interoperability for Microwave Access (WiMAX) and Long Term Evolution (LTE). Resource allocation mechanisms in LTE are very critical issues, because scheduling algorithms have the main responsibility for determining how to allocate radio resources for different users. In this paper a dynamically adapted Proportional Fair (PF) scheduling algorithm for capacity enhancement of LTE system is proposed. Performance comparison with the conventional PF downlink scheduler, which is characterized by high fairness but with low throughput, and the Best-Channel Quality Indicator( Best-CQI) scheduling algorithm which is characterized by high throughput but with poor fairness performance is presented. Simulation results show that the proposed algorithm enhances the overall system capacity and also provides fairness in the distribution of the resources. The proposed algorithm improves the average cell throughput by more than 31 %, with a slight degradation in the fairness level as compared with the conventional Proportional Fair PF scheduling algorithm.

Keywords: LTE, packet scheduling, PF, Fairness, OFDM.

Received April 24, 2015; accepted March 13, 2016


Sunday, 26 August 2018 03:48

Temporal Tracking on Videos with Direction Detection

Shajeena Johnson

Department of Computer Science and Engineering, James College of Engineering and Technology, India

Abstract: Tracking is essentially a matching problem. This paper proposes a tracking scheme for video objects on compressed domain. This method mainly focuses on locating the object region and predicting (evolving) the detection of movement, which improves tracking precision. Motion Vectors (MVs) are used for block matching. At each frame, the decision of whether a particular block belongs to the object being tracked is made with the help of histogram matching. During the process of matching and evolving the direction of movement, similarities of target region are compared to ensure that there is no overlapping and tracking performed in a right way. Experiments using the proposed tracker on videos demonstrate that the method can reliably locate the object of interest effectively.

Keywords: Motion vector, distance measure, histogram, block matching, DCT, tracking. 

Received August 19, 2014; accepted April 2, 2015

Sunday, 26 August 2018 03:37

Medical Image Segmentation Based on Fuzzy Controlled Level Set and Local Statistical Constraints

Mohamed Benzian1,2 and Nacéra Benamrane2

1Département d'Informatique, Université Abou Bekr Belkaid-Tlemcen, Algérie

2Laboratoire SIMPA, Département d'Informatique, Université des Sciences et de la Technologie d'Oran Mohamed Boudiaf, Algérie

Abstract: Image segmentation is one of the most important fields in artificial vision due to its complexity and the diversity of its application to different image cases. In this paper, a new Region of Interest (ROI) segmentation in medical images approach is proposed, based on modified level sets controlled by fuzzy rules and incorporating local statistical constraints (mean, variance) in level set evolution function, and low image resolution analysis by estimating statistical constraints and curvature of curve at low image scale. The image and curve at low resolution provide information on rough variation of respectively image intensity and curvature value. The weights of different constraints are controlled and adapted by fuzzy rules which regularize their influence. The objective of using low resolution image analysis is to avoid stopping the evolution of the level set curve at local maxima or minima of images. This method is tested on medical images. The obtained results of the technique presented are satisfying and give a good precision.

Keywords: Segmentation, level sets, medical images, image resolution, fuzzy rules, ROI.

Received April 8, 2015; accepted December 23, 2015

Sunday, 26 August 2018 03:28

Phishing Detection using RDF and Random Forests

Vamsee Muppavarapu, Archanaa Rajendran, and Shriram Vasudevan

Department of Computer Science and Engineering, Amrita Vishwa Vidyapeetham University, India

Abstract: Phishing is one of the major threats in this internet era. Phishing is a smart process where a legitimate website is cloned and victims are lured to the fake website to provide their personal as well as confidential information, sometimes it proves to be costly. Though most of the websites will give a disclaimer warning to the users about phishing, users tend to neglect it. It is not a fully responsible action by the websites also and there is not much that the websites could really do about it. Since phishing has been in persistence for a long time, many approaches have been proposed in past that can detect phishing websites but very few or none of them detect the target websites for these phishing attacks, accurately. Our proposed method is novel and an extension to our previous work, where we identify phishing websites using a combined approach by constructing Resource Description Framework (RDF) models and using ensemble learning algorithms for the classification of websites. Our approach uses supervised learning techniques to train our system. This approach has a promising true positive rate of 98.8%, which is definitely appreciable. As we have used random forest classifier that can handle missing values in dataset, we were able to reduce the false positive rate of the system to an extent of 1.5%. As our system explores the strength of RDF and ensemble learning methods and both these approaches work hand in hand, a highly promising accuracy rate of 98.68% is achieved.

Keywords: Phishing, ensemble learning, RDF models, phishing target, metadata, vocabulary, random forests.

Received April 22, 2015; accepted September 20, 2015

Sunday, 26 August 2018 03:25

Multi-Sensor Fusion based on DWT, Fuzzy

Histogram Equalization for Video Sequence

Nada Habeeb1, Saad Hasson2, and Phil Picton3

1Technical College of Management, Middle Technical University, Iraq

2Department of Computer Science, College of Science, University of Babylon, Iraq

3School of science and Technology, University of Northampton, UK

Abstract: Multi-sensor fusion is a process which combines two or more sensor datasets of same scene resulting a single output containing all relevant information. The fusion process can work in the spatial domain and the transform domain. The spatial domain fusion methods are easy to implement and have low computational complexity, but they may produce blocking artefacts and out of focus which means that the fused image got blur. In this paper, fusion algorithm has been proposed to solve this problem based on Discrete Wavelet Transform (DWT), Fuzzy Histogram Equalization, and De-blurring Kernel. In addition, two fusion techniques: Maximum selection and weighted average were developed based on Mean statistical technique. The performance of the proposed method has been tested on the real and synthetic datasets. Experimental results showed the proposed fusion method with traditional and developed fusion rules gives improvement in fused results.

Keywords: Multi-sensor fusion, discrete wavelet transform, fuzzy histogram equalization, de-blurring kernels, principle component analysis.

Received September 4, 2015; accepted December 27, 2015

Full text 

Sunday, 26 August 2018 03:23

Maximum Spanning Tree Based Redundancy

Elimination for Feature Selection of High

Dimensional Data

Bharat Singh and Om Prakash Vyas

Department of Information Technology, Indian Institute of Information Technology-Allahabad, India

Abstract: Feature selection adheres to the phenomena of preprocessing step for High Dimensional data to obtain optimal results with reference of speed and time. It is a technique by which most prominent features can be selected from a set of features that are prone to contain redundant and relevant features. It also helps to lighten the burden on classification techniques, thus makes it faster and efficient.We introduce a novel two tiered architecture of feature selection that can able to filter relevant as well as redundant features. Our approach utilizes the peculiar advantage of identifying highly correlated nodes in a tree. More specifically, the reduced dataset comprises of these selected features. Finally, the reduced dataset is tested with various classification techniques to evaluate their performance. To prove its correctness we have used many basic algorithms of classification to highlight the benefits of our approach. In this journey of work we have used benchmark datasets to prove the worthiness of our approach.

Keywords: Data mining, feature selection, tree based approaches, maximum spanning tree, high dimensional data.

Received February 15, 2015; accepted December 21, 2015

 Full text

Sunday, 26 August 2018 03:21

Auto-Poietic Algorithm for Multiple Sequence Alignment

Amouda Venkatesan and Buvaneswari Shanmugham

Centre for Bioinformatics, Pondicherry University, India

Abstract: The concept of self-organization is applied to the operators and parameters of genetic algorithm to develop a novel Auto-poietic algorithm solving a biological problem, Multiple Sequence Alignment (MSA). The self-organizing crossover operator of the developed algorithm undergoes a swap and shuffle process to alter the genes of chromosomes in order to produce better combinations. Unlike Standard Genetic Algorithms (SGA), the mutation rate of auto-poietic algorithm is not fixed. The mutation rate varies cyclically based on the improvement of fitness value in turn, determines the termination point of algorithm. Automated assignment of various parameter values reduces the intervention and inappropriate settings of parameters from user without prior the knowledge of input. As an advantage, the proposed algorithm also circumvents the major issues in standard genetic algorithm, premature convergence and time requirements to optimize the parameters. Using Benchmark Alignment Database (BAliBASE) reference multiple sequence alignments, the efficiency of the auto-poietic algorithm is analyzed. It is evident that the performance of auto-poietic algorithm is better than SGA and produces better alignments compared to other MSA tools.

Keywords: Auto-poietic, crossover, genetic algorithm, mutation, multiple sequence alignment, selection.

Received October 27, 2014; accepted November 29, 2015

Sunday, 26 August 2018 03:16

A Novel Architecture of Medical Image Fusion

Based on YCbCr-DWT Transform

Behzad Nobariyan1, Nasrin Amini2, Sabalan Daneshvar3, and Ataollah Abbasi4

1Faculty of Electrical Engineering, Sahand University of Technology, Iran

2Faculty of Biomedical Engineering, Islamic Azad University Branch of Science and Research, Iran

3Faculty of Electrical and Computer Engineering, University of Tabriz, Iran

4Faculty of Electrical Engineering, Sahand University of Technology, Iran

Abstract: Image fusion is one of the most modern, accurate and useful diagnostic techniques in medical imaging. Mainly, image fusion tries to offer a method for solving the problem that no system is able to integrate functional and anatomical information. Multiple image fusion of brain is very important for clinical applications. Positron Emission Tomography (PET) image indicates the brain function and Single-Photon Emission Computerized Tomography (SPECT) indicates local performance in the internal organs like heart and brain imaging. Both of these images are multi-spectral images and have a low spatial resolution. The Magnetic Resonance Imaging (MRI) image shows the brain tissue anatomy and contains no functional information. A good fusion scheme should preserve the spectral characteristics of the source multispectral image as well as the high spatial resolution characteristics of the source panchromatic image. There are many methods for image fusion but each of them has certain limitations. The studies have shown that YCbCr preserves spatial information and Discrete Wavelet Transforms (DWT) preserves spectral information without distortion. The proposed method contains the advantages of both methods and it preserves spatial and spectral information without distortion. Visual and statistical analyses show that the results of our algorithm considerably enhance the fusion quality in connection with: discrepancy, average gradient and Mutual information; compared to fusion methods including, Hue-Intensity-Saturation (HIS), YCbCr, Brovey, Laplacian-pyramid, Contourlet and DWT.

Keywords: YCbCr, DWT, PET, SPECT, image fusion.

 

Received April 24, 2015; accepted March 9, 2016

Full text
Sunday, 26 August 2018 03:13

Reverse Engineering of Object Oriented System using Hierarchical Clustering

Aman Jatain1 and Deepti Gaur2

 1Department of Computer Science, Amity University, India

2Department of Computer Science, North Cap University, India

Abstract: Now a day’s common problem faced by software community is to understand the legacy code. A decade ago the legacy code referred as the code written in language like Common Business Oriented Language (COBOL) or Formula Translation (FORTRAN). Today software engineers primarily use object oriented language like C++ and Java. This implies that tomorrow’s legacy code is written today because object oriented programs are even more difficult and complex to understand which leads us towards making software that is vague and having insufficient design documentation. Object oriented programming produce many problems to software developers in maintenance phase. So reverse engineering methodologies can be applied to resolve it. In literature various techniques has been proposed by researchers to recover the architecture and components of legacy systems. The use of clustering algorithms has recently been discussed by many for reverse engineering and architecture recovery. Methodology: In this paper Rational Software Architect (RSA) is used to recover the design from source code during reverse engineering process and then feature selection method is applied to select the features of software system. Hierarchical clustering is used after calculating the similarity measure between classes to cluster the similar classes into one component. The proposed technique is demonstrated by a case study.

Keywords: Clustering, feature selection, hierarchical, reverse engineering, rational software architect.

Received April 28, 2015; accepted November 29, 2015

Sunday, 26 August 2018 03:09

Enhanced Hybrid Prediction Models for Time Series Prediction

Purwanto Purwanto1 and Chikkannan Eswaran2

1Faculty of Computer Science, Dian Nuswantoro University, Indonesia

2Faculty of Computing and Informatics, Multimedia University, Malaysia

Abstract: Statistical techniques have disadvantages in handling the non-linear pattern. Soft Computing (SC) techniques such as artificial neural networks are considered to be better for prediction of data with non-linear patterns. In the real-life, time-series data comprise complex pattern, and hence it may be difficult to obtain high prediction accuracy rates using the statistical or SC techniques individually. We propose two enhanced hybrid models for time series prediction. The first model is an enhanced hybrid model combining statistical and neural network techniques. Using this model, one can select the best statistical technique as well as the best configuration for the neural network for time series prediction. The second model is an enhanced adaptive neuro-fuzzy inference system which combines fuzzy inference system and neural network. The proposed enhanced Adaptive Neuro-Fuzzy Inference Systems (ANFIS) model can determine the optimum input lags for obtaining the best accuracy results. The prediction accuracies of the two proposed hybrid models are compared with those obtained with other models based on three time series data sets. The results indicate that the proposed hybrid models yield better accuracy results compared to Autoregressive Integrated Moving Average (ARIMA), exponential smoothing, moving average, weighted moving average and Neural Network models.

Keywords: Hybrid model, adaptive neuro-fuzzy inference systems, soft computing, neural network, statistical techniques.

Received March 25, 2015; accepted October 7, 2015

Full text  

 
Sunday, 26 August 2018 03:07

Image Steganography Based on

Hamming Code and Edge Detection

Shuliang Sun

School of Electronics and Information Engineering, Fuqing Branch of Fujian Normal University, China

Abstract: In this paper a novel algorithm which is based on hamming code and 2k correction is proposed. The new method also utilizes canny edge detection and coherent bit length. Firstly, canny edge detector is applied to detect the edge of cover image and only edge pixels are selected for embedding payload. In order to enhance security, the edge pixels are scrambled. Then hamming encoding is practiced to code the secret data before embedded. Calculate coherent bit length L on the base of relevant edge pixels and replace with L bits of payload message. Finally, the method of 2k correction is applied to achieve better imperceptibility in stego image. The experiment shows that the proposed method is more advantage in Peak Signal-to-Noise Ratio (PSNR), capacity and universal image quality index (Q) than other methods.

Keywords: Hamming code, 2k correction, coherent bit length, canny edge detector.

Received October 5, 2015; accepted March 30, 2016

Full text  


  
Sunday, 26 August 2018 02:56

A New Method for Curvilinear Text line Extraction and Straightening of Arabic Handwritten Text

Ayman Al Dmour1, Ibrahim El rube'2, and Laiali Almazaydeh1

1Faculty of Information Technology, Al-Hussein Bin Talal University, Jordan

2Department of Computer Engineering, Taif University, KSA

Abstract: Line extraction is a critical step from one of the main subtasks of Document Image Analysis, which is layout analysis. This paper presents a new method for curvilinear text line extraction and straightening in Arabic handwritten documents. The proposed method is based on a strategy that consists of two distinct steps. First, text line is extracted based on morphological dilation operation. Secondly, the extracted text line is straighten in two sub-steps: Course tuning of text line orientation based on Hough transform, then fine tuning based on centroid alignment of the connected component that forms the text line. The proposed approach has been extensively experimented on samples from the benchmark datasets of KFUPM Handwritten Arabic TexT (KHATT) and Arabic Handwriting DataBase (AHDB). Experimental results show that, the proposed method is capable of detecting and straightening curvilinear text lines even on challenging Arabic handwritten documents.

Keywords: Document image analysis, arabic handwriting, text line extraction, hough transform.

Received January 14, 2016; accepted May 11, 2016

Full text  

 
Sunday, 26 August 2018 02:46

Transfer-based Arabic to English Noun Sentence

Translation Using Shallow Segmentation

Namiq Abdullah

Department of Electrical and Computer Engineering, University of Duhok, Iraq

Abstract: The quality of machine translation systems decreases considerably when dealing with long sentences. In this paper, a transfer-based system is developed for translating long Arabic noun sentences into English. A simple method used for dividing a long sentence into phrases based on conjunctions, prepositions, and quantifier particles. These particles divide a sentence into phrases. The phrases of a source sentence are translated individually. In the end of translation process, target sentence is constructed by connecting the translated phrases. The system was tested on 100 thesis long titles from the management and economy domain. The results show that the method is very efficient with most of the tested sentences.

Keywords: Machine translation, transfer-based approach, noun phrases, sentence partitioning.

Received May 15, 2015; accepted March 24, 2016

Full text  


  
Sunday, 26 August 2018 02:43

A Network Performance Aware QoS Based

Workflow Scheduling for Grid Services

 Shinu John1 and Maluk Mohamed2

 1Department of Computer Science and Engineering, St Thomas College of Engineering and Technology, India

2Department of Computer Science and Engineering, MAM College of Engineering and Technology, India

Abstract: Grids enable sharing, selection and aggregation of geographically distributed resources among various organizations. They are now emerging as promising computing paradigms for resource and compute intensive scientific workflow applications modeled as a Directed Acyclic Graph (DAG) with intricate inter-task dependencies. Job scheduling is an important and challenging issue in a grid environment. There are various scheduling algorithm proposed for grid environments to distribute the load among processors and maximize resource utilization while reducing task execution time. Task execution time is not the only parameter to be improved; various Quality of Service (QoS) parameters are also to be considered in job scheduling in grid computing. In this Research we have studied the existing QoS based Task scheduling, work flow scheduling and formulated the problem. The possible solutions are developed for the problems identified in existing algorithms. The scheduling of dependent task (work flow) is more challenging than independent task scheduling. The scheduling of both dependent and independent tasks with satisfying QOS requirements of users is a very challenging issue in grid computing. This paper proposes a Novel Network aware QoS workflow scheduling method for Grid Services. The proposed scheduling algorithm considers network and QoS constraints. The goal of the proposed scheduling algorithm is to implement the workflow schedule so that it reduces execution time and resource cost and yet meets the deadline imposed by the user. The experimental result shows that the proposed algorithm improves the success ratio of tasks and throughput of resources while reducing makespan and workflow execution cost.

Keywords: Grid scheduling, QoS, DAG, execution time, deadline, trust rate.

Received June 25, 2014; accepted September 7, 2015

Full text   

Sunday, 26 August 2018 02:36

Hyperspectral Image Segmentation Based on Enhanced Estimation of Centroid with Fast K-Means

Saravana Kumar Veligandan1 and Naganathan Rengasari2

1Department of Information Technology, SreeNidhi Institute of Science and Technology, India

2Symbiosis Institute of Computer Studies and Research, Symbiosis International (Deemed University), India

Abstract: In this paper, the segmentation process is observant on hyperspectral satellite images. A novel approach, hyperspectral image segmentation based on enhanced estimation of centroid with unsupervised clusters such as fast k-means, fast k-means (weight), and fast k-means (careful seeding) has been addressed. Besides, a cohesive image segmentation approach based on inter-band clustering and intra-band clustering is processed. Moreover, the inter band clustering is accomplished by above clustering algorithms, while the intra band clustering is effectuated using Particle Swarm Clustering algorithm (PSC) with Enhanced Estimation of Centroid (EEOC). The hyperspectral bands are clustered and a single band which has a paramount variance from each cluster is opting for. This constructs the diminished set of bands. Finally, PSC EEOC carried out the segmentation process on the diminished bands. In addition, we compare the result produce in these methods by statistical analysis based on number of pixel, fitness value, and elapsed time.

Keywords: Fast k-means, fast k-mean (weight), fast k- means (careful seeding), and particle swarm clustering algorithm.

 

Received March 16, 2015; accepted September 7, 2015
 
Full text  
Sunday, 26 August 2018 02:33

Multi-Classifier Model for Software Fault Prediction

Pradeep Singh1 and Shrish Verma2

1Department of Computer Science and Engineering, National Institute of Technology, Raipur

2Department of Electronics and Telecommunication Engineering, National Institute of Technology, Raipur

Abstract: Prediction of fault prone module prior to testing is an emerging activity for software organizations to allocate targeted resource for development of reliable software. These software fault prediction depend on the quality of fault and related code extracted from previous versions of software. This paper, presents a novel framework by combining multiple expert machine learning systems. The proposed multi-classifier model takes the benefits of best classifiers in deciding the faulty modules of software system with consensus prior to testing. An experimental comparison is performed with various outperformer classifiers in the area of fault prediction. We evaluate our approach on 16 public dataset from promise repository which consists of National Aeronautics and Space Administration( NASA) Metric Data Program (MDP) projects and Turkish software projects. The experimental result shows that our multi classifier approach which is the combination of Support Vector Machine (SVM), Naive Bayes (NB) and Random forest machine significantly improves the performance of software fault prediction.

Keywords: Software metrics, software fault prediction, machine learning.

Received February 7, 2015; accepted September 7, 2015

Full text  

Sunday, 26 August 2018 02:31

Traceability between Code and Design Documentation in Database Management System: A Case Study

Mohammed Akour, Ahmad Saifan, and Osama Ratha'an

Computer Information Systems Department, Yarmouk University, Jordan

Abstract: Traceability builds many strong connections or links between requirements and design, so the main purpose of traceability is to maintain consistency between a high level conceptual view and a low level implementation view. The purpose of this paper is to have full consistency between all components over all phases in the oracle designer tool by allowing traceability to be carried out not only between the requirements and design but also between the code and design. In this paper, we propose a new methodology to support traceability and completeness checking between code and design of oracle database applications. The new algorithm consists of a set of interrelated steps to initialize the comparison environment. An example of a student information System is used to illustrate the work.

Keywords: Traceability, oracle designer, completeness checking, design, source code, database, pl/sql, testing.

Received February 18, 2016; accepted June 26, 2016
 
Sunday, 26 August 2018 02:29

Edge Preserving Image Segmentation using Spatially Constrained EM Algorithm

Meena Ramasamy1 and Shantha Ramapackiam2

1Department of Electronics and Communication Engineering, Sethu Institute of Technology, India

2Department of Electronics and Communication Engineering, Mepco Schlenk Engineering College, India

Abstract: In this paper, a new method for edge preserving image segmentation based on the Gaussian Mixture Model (GMM) is presented. The standard GMM considers each pixel as independent and does not incorporate the spatial relationship among the neighboring pixels. Hence segmentation is highly sensitive to noise. Traditional smoothing filters average the noise, but fail to preserve the edges. In the proposed method, a bilateral filter which employs two filters - domain filter and range filter, is applied to the image for edge preserving smoothing. Secondly, in the Expectation Maximization algorithm used to estimate the parameters of GMM, the posterior probability is weighted with the Gaussian kernel to incorporate the spatial relationship among the neighboring pixels. Thirdly, as an outcome of the proposed method, edge detection is also done on images with noise. Experimental results obtained by applying the proposed method on synthetic images and simulated brain images demonstrate the improved robustness and effectiveness of the method.

Keywords: Gaussian mixture model, expectation maximization, bilateral filter, image segmentation.

Received December 23, 2014; accepted December 21, 2015

Full text

 
Sunday, 26 August 2018 02:27

Paradigma: A Distributed Framework for Parallel Programming

Sofien Gannouni, Ameur Touir, and Hassan Mathkour

College of Computer and Information Sciences, King Saud University, Saudi Arabia

Abstract: Recent advances in high-speed networks and the newfound ubiquity of powerful processors have revolutionized the nature of parallel computing. It is becoming increasingly attractive to perform parallel tasks on distant, autonomous, and heterogeneous networked machines. This paper presents a simple and efficient new distributed framework for parallel programming known as Paradigma. In this framework, parallel program development is simplified using the Gamma formalism, providing sequential programmers with a straightforward mechanism for solving large-scale problems in parallel. The programmer simply specifies the action to be performed on an atomic data element known as a molecule. The workers compete in simultaneously running the action specified on the various molecules extracted from the input until the entire dataset is processed. The proposed framework is dedicated for fine-grained parallel processing and supports both the Simple program multiple data and multiple program multiple data programming models.

Keywords: Distributed systems, parallel programming, gamma formalism, single program multiple data, multiple program multiple data.

Received March 5, 2015; accepted March 9, 2016

Full text

 
Top
We use cookies to improve our website. By continuing to use this website, you are giving consent to cookies being used. More details…