Qualification: 
Ph.D
sajeevgp@am.amrita.ac.in

Dr. Sajeev currently serves as Associate Professor at Department of Computer Science and Engineering, School of Engineering, Amritapuri Campus. He received his Ph.D. from the National Institute of Technology Calicut, India. His research interests include Open Systems, Web Cache Systems, P2P Networking, Traffic Measurement & Modeling and Web Server Acceleration, and has published many research papers in international journals and conferences. He is the reviewer of International Journal of Computers & Electrical Engineering, Elsevier and the TPC Member of IEEE organised  conferences. 

Publications

Publication Type: Conference Paper

Year of Publication Publication Type Title

2017

Conference Paper

P. Devika, Jisha R. C., and Dr. Sajeev G. P., “A novel approach for book recommendation systems”, in 2016 IEEE International Conference on Computational Intelligence and Computing Research, ICCIC 2016, Chennai, India, 2017.[Abstract]


Recommendation systems are widely used in ecommerce applications. A recommendation system intends to recommend the items or products to a particular user, based on user's interests, other user's preferences, and their ratings. To provide a better recommendation system, it is necessary to generate associations among products. Since e-commerce and social networking sites generates massive data, traditional data mining approaches perform poorly. Also, the pattern mining algorithm such as the traditional Apriori suffers from high latency in scanning the large database for generating association rules. In this paper we propose a novel pattern mining algorithm called as Frequent Pattern Intersect algorithm (FPIntersect algorithm), which overcomes the drawback of Apriori. The proposed method is validated through simulations, and the results are promising.

More »»

2017

Conference Paper

K. Pavani and Dr. Sajeev G. P., “A Novel Web Crawling Method for Vertical Search Engines”, in 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2017.[Abstract]


The main goal of focused web crawlers is to retrieve as many relevant pages as possible. However, most of the crawlers use page rank algorithm to lineup the pages in the crawler frontier. Since the page rank algorithm suffers from the drawback of “Richer get rich phenomenon”, focused crawlers often fail to retrieve the hidden relevant pages. This paper presents a novel approach for retrieving the hidden and relevant pages by combining rank and semantic similarity information. The model is validated by crawling the real web with different topics and the results are promising.

More »»

2017

Conference Paper

P. Das, Jisha, R. C., and Dr. Sajeev G. P., “Adaptive Web Personalization System Using Splay Tree”, in 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2017.[Abstract]


Web personalization helps in understanding the user interests and creating customized experiences for users. However the user preferences changes dynamically over a period. In order to adapt with the changing information needs of the user, we have developed a novel web personalization system that captures the user changing interest by analyzing the timing information. We use splay tree, which is a self-adaptive data structure, for tracking the changing trends of the users. The proposed web personalization model is validated by building a simulation model, with real and synthetic dataset, and the quality of results are promising.

More »»

2017

Conference Paper

A. R. Kurup and Dr. Sajeev G. P., “Task recommendation in reward-based crowdsourcing systems”, in 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2017.[Abstract]


Crowdsourcing systems are distributed problem solving platforms, where small tasks are channelled to a crowd in the form of open calls for solutions. Reward based crowdsourcing systems tries to attract the interested and capable workers to provide solutions in return for monetary rewards. We study the task recommendation problem in reward based crowdsourcing platforms, where we leverage both implicit and explicit features of the worker-reward and worker-task attributes. Given a set of workers, set of tasks, participation, winner attributes, we intend to recommend tasks to workers by exploiting interactions between tasks and workers. Two models based on worker-reward based features and worker task based features are presented. The proposed approach is compared with multiple related techniques using real world dataset.

More »»

2016

Conference Paper

Dr. Sajeev G. P. and Nair, L. M., “LASER: A novel hybrid peer to peer network traffic classification technique”, in 2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2016.[Abstract]


The popularity of Peer-to-peer (P2P) applications have shown a massive growth in recent times, and P2P traffic contributes considerably to the today's internet traffic. For efficient network traffic management and effective malware detection, P2P traffic classification is indispensable. This paper proposes LASER, Longest Common Subsequence (LCS)-based Application Signature ExtRaction technique, algorithm, a novel hybrid network traffic classification technique which classifies the P2P traffic into malicious P2P and non-malicious P2P traffic. The proposed classifier analyzes the header information for creating a communication module. Further, the signature is extracted from the payload information. We build the classifier by aggregating the information of header and the payload. The proposed hybrid classifier is analyzed for its performance and the results are promising. More »»

2016

Conference Paper

T. R. Krishnaprasad and Dr. Sajeev G. P., “A novel method for resource discovery from a range based request in a P2P network”, in 2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2016.[Abstract]


Peer to Peer (P2P) systems have increased the curiosity and pathways for people to discover and share the resources. While various methods have been proposed in the discovery of discrete value based resources, there is also a surging interest in discovering a range of resources for a given request. This work is a novel design of a P2P network that adheres to range requests and seeks to discover the resources sought for in the request. The proposed model seeks to find out the range of resource values from within a P2P network of nodes that are in a circular overlay structure. The validation of the design reaches the conclusion that the proposed model increases in efficiency as the number of hubs increases with respect to discovering a range of resources in the hubs. More »»

2016

Conference Paper

Dr. Sajeev G. P. and Ramya, P. T., “Effective web personalization system based on time and semantic relatedness”, in 2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2016.[Abstract]


The key aspect in building a Web personalization system is the user's navigational pattern. However, the navigational pattern alone is insufficient to capture the user's interest and behavior. This paper proposes a novel web personalization system that accepts the timing information, semantic information along with the navigational pattern, and classifies the users according their interest and behavior on the site. The proposed model is validated by constructing a Web personalization model using the real and synthetic data and the results are promising.

More »»

2016

Conference Paper

M. Annam and Dr. Sajeev G. P., “Entropy based informative content density approach for efficient web content extraction”, in 2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2016.[Abstract]


Web content extraction is a popular technique for extracting the main content from web pages and discards the irrelevant content. Extracting only the relevant content is a challenging task since it is difficult to determine which part of the web page is relevant and which part is not. Among the existing web content extraction methods, density based content extraction is one popular method. However density based methods, suffer from poor efficiency, especially when the pages containing less information and long noise. We propose a web content extraction technique build on Entropy based Informative Content Density algorithm (EICD). The proposed EICD algorithm initially analyses higher text density content. Further, the entropy-based analysis is performed for selected features. The key idea of EICD is to utilize the information entropy for representing the knowledge that correlates to the amount of informative content in a page. The proposed method is validated through simulation and the results are promising. More »»

2015

Conference Paper

P. T. Ramya and Dr. Sajeev G. P., “Building Web Personalization System with Time-Driven Web Usage Mining”, in Proceedings of the Third International Symposium on Women in Computing and Informatics, New York, NY, USA, 2015.[Abstract]


Web personalization is a powerful tool used for personalizing the Websites. The personalization system aims at suggesting the Web pages to the users based on their navigational patterns. Use of attributes such as time, popularity of Web objects makes the model more efficient. This paper proposes a novel Web personalization model which utilizes time attributes, such as duration of visit, inter-visiting time, burst of visit, and the user's navigational pattern. Test results indicate that the proposed model explores the user's behaviour and their interest. More »»

2015

Conference Paper

L. M. Nair and Dr. Sajeev G. P., “Internet Traffic Classification by Aggregating Correlated Decision Tree Classifier”, in Proceedings of the 2015 Seventh International Conference on Computational Intelligence, Modelling and Simulation, 2015.[Abstract]


Peer-to-Peer (P2P) traffic shows a rapid growth in recent times. For efficient malware detection and network traffic management P2P network traffic classification is essential. The existing P2P traffic classification methods includes port-based, signature-based, pattern-based, and statistics based methods. However, none of these methods proved to be effective for the traffic classification in terms of the classification accuracy. This paper proposes a novel classification technique which classifies the internet traffic into P2P and non-P2P traffic with more accuracy and less computational overhead. The proposed classifier is the flow based classifier, that analyses the behavioural patterns utilizing the correlation metric algorithm. The proposed classifier is analyzed for its performance and the results are encouraging. More »»

2015

Conference Paper

Dr. Sajeev G. P. and M, R. Chandran, “Intelligent Pollution Controlling Mechanism for Peer to Peer Caches”, in Proceedings of the 2015 Seventh International Conference on Computational Intelligence, Modelling and Simulation, 2015.[Abstract]


This paper addresses the problem of cache pollution in P2P (Peer-to-peer) systems. P2P traffic has a significant impact on ISPs as it accounts for more than half of the all traffic. This will cause some negative impact on the Internet like network congestion, high latency. P2P caching is an efficient method for handling this problem. Most of the P2P cache systems suffer from Cache Pollution. This research proposes an intelligent cache updation algorithm, which helps in reducing the cache pollution by using the probabilistic approach. The proposed method is evaluated in comparison with Least Frequently Used and Least Recently Used algorithms. More »»

2011

Conference Paper

Dr. Sajeev G. P. and Sebastian, M. P., “Comparing the Performance of Multinomial Logistic Regression and Neural Network Models in Web Cache Content Classification”, in ICMLC 2011: the 3rd International Conference on Machine Learning and Computing, 2011.

2010

Conference Paper

Dr. Sajeev G. P. and Sebastian, M. P., “Building a semi intelligent web cache with light weight machine learning”, in Intelligent Systems (IS), 2010 5th IEEE International Conference, 2010.[Abstract]


This paper proposes a novel admission and replacement technique for web caching, which utilizes the multinomial logistic regression (MLR) as classifier. The MLR model is trained for classifying the web cache's object worthiness. The parameter object worthiness is a polytomous (discrete) variable which depends on the traffic and the object properties. Using worthiness as a key, an adaptive caching model is proposed. Trace driven simulations are used to evaluate the performance of the scheme. Test results show that a properly trained MLR model yields good cache performance in terms of hit ratios and disk space utilization, making the proposed scheme as a viable semi intelligent caching scheme. More »»

2010

Conference Paper

Dr. Sajeev G. P. and Sebastian, M. P., “A scheme for adaptive web caching based on multi level object classification”, in Intelligent and Advanced Systems (ICIAS), 2010 International Conference on, 2010.[Abstract]


Multi-level classification of web objects in caching is relatively an unexplored area. This paper proposes a novel caching scheme which utilizes a multi-level class information. A MLR (Multinomial Logistics Regression) based classifier is constructed using the information from web logs. Simulation results confirm that the model has good prediction capability and suggest that the proposed approach can improve the performance of the cache substantially. More »»

2005

Conference Paper

Dr. Sajeev G. P. and Sebastian, M. P., “Popularity-Aware Renewal of Dynamic Web Documents with $\$Zipf$\$ Law”, in ADCOM'05: the 13th International Conference on Advances in Computers and Communications, 2005.

Publication Type: Journal Article

Year of Publication Publication Type Title

2013

Journal Article

Dr. Sajeev G. P. and Sebastian, M. P., “Building semi-intelligent web cache systems with lightweight machine learning techniques”, Computers & Electrical Engineering, vol. 39, pp. 1174–1191, 2013.[Abstract]


This paper proposes a novel admission and replacement technique for web caching, which utilizes the multinomial logistic regression (MLR) as classifier. The MLR model is trained for classifying the web cache's object worthiness. The parameter object worthiness is a polytomous (discrete) variable which depends on the traffic and the object properties. Using worthiness as a key, an adaptive caching model is proposed. Trace driven simulations are used to evaluate the performance of the scheme. Test results show that a properly trained MLR model yields good cache performance in terms of hit ratios and disk space utilization, making the proposed scheme as a viable semi intelligent caching scheme.

More »»

2010

Journal Article

Dr. Sajeev G. P. and Sebastian, M. P., “A novel content classification scheme for web caches”, Evolving Systems, vol. 2, pp. 101–118, 2010.[Abstract]


Web caches are useful in reducing the user perceived latencies and web traffic congestion. Multi-level classification of web objects in caching is relatively an unexplored area. This paper proposes a novel classification scheme for web cache objects which utilizes a multinomial logistic regression (MLR) technique. The MLR model is trained to classify web objects using the information extracted from web logs. We introduce a novel grading parameter worthiness as a key for the object classification. Simulations are carried out with the datasets generated from real world trace files using the classifier in Least Recently Used-Class Based (LRU-C) and Least Recently Used-Multilevel Classes (LRU-M) cache models. Test results confirm that the proposed model has good online learning and prediction capability and suggest that the proposed approach is applicable to adaptive caching. More »»

2009

Journal Article

Dr. Sajeev G. P. and Sebastian, M. P., “Analyzing the Long Range Dependence and Object Popularity in Evaluating the Performance of Web Caching”, International Journal of Information Technology and Web Engineering (IJITWE), vol. 4, pp. 25–37, 2009.[Abstract]


Web cache systems enhance Web services by reducing the client side latency. To deploy an effective Web cache, analysis of the traffic characteristics is indispensable. Various reported results of traffic analysis show evidences of long range dependence (LRD) in the data stream and rank distribution of the documents in Web traffic. This article analyzes Web cache traffic properties like LRD and rank distribution based on the traces collected from NLANR (National Laboratory of Applied Network Research) cache servers. Traces are processed to investigate the performance of Web cache servers and traffic patterns. Statistical tools are utilized to measure the strengths of the LRD and popularity. The Hurst parameter, which is a measure of the LRD, is estimated using various statistical methods. It is observed that the presence of LRD in the traffic is feeble and does not influence the Web cache performance. More »»

Publication Type: Book Chapter

Year of Publication Publication Type Title

2011

Book Chapter

Dr. Sajeev G. P. and Sebastian, M. P., “Analyzing the Traffic Characteristics for Evaluating the Performance of Web Caching”, in Web Engineered Applications for Evolving Organizations: Emerging Knowledge, vol. 1, Information Science Reference, 2011, pp. 196–208.[Abstract]


Web cache systems enhance Web services by reducing the client side latency. To deploy an effective Web cache, study about traffic characteristics is indispensable. Various reported results show the evidences of long range dependence (LRD) in the data stream and rank distribution of the documents in Web traffic. This chapter analyzes Web cache traffic properties such as LRD and rank distribution based on the traces collected from NLANR (National Laboratory of Applied Network Research) cache servers. Traces are processed to investigate the performance of Web cache servers and traffic patterns. Statistical tools are utilized to measure the strengths of the LRD and popularity. The Hurst parameter, which is a measure of the LRD is estimated using various statistical methods. It is observed that presence of LRD in the trace is feeble and has practically no influence on the Web cache performance. More »»

207
PROGRAMS
OFFERED
5
AMRITA
CAMPUSES
15
CONSTITUENT
SCHOOLS
A
GRADE BY
NAAC, MHRD
8th
RANK(INDIA):
NIRF 2018
150+
INTERNATIONAL
PARTNERS