Refine Search

New Search

Results in Journal JUTI: Jurnal Ilmiah Teknologi Informasi: 337

(searched for: journal_id:(1349332))
Page of 7
Articles per Page
by
Show export options
  Select all
Dino Budi Prakoso, Royyana Muslim Ijtihadie, Tohari Ahmad
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 19; https://doi.org/10.12962/j24068535.v19i1.a1009

Abstract:
In the technology world especially in the field of current network of Autonomous Systems connectivity (AS) is indispensable. Especially against the dynamic routing protocols that are often used compared to static routing protocols. In supporting this current network, it takes efficient and effective routing protocols capable of covering a sizable scale. Software Defined Network (SDN) is a technological innovation in the network world that has a separate Control Plane and Data Plane that makes it easy to configure on the Control Plane side. Control Plane is the focal point on a process of bottleneck in SDN architecture. Performance is a critical issue in large-scale network implementations because of the large demand load occurring in the Control Plane by generating low throughput value. This research will be conducted testing on the Hybrid network of SDN by using OSPF routing protocol, based on the Fibbing architecture implemented on the system network Hybrid SDN also able to assist in improving performance, but there are constraints when sending flooding which is used as a fake node forming. Many nodes are not skipped as distribution lines in the formation of a fake node, in which case it will certainly affect the value of throughput to be unstable and decrease. This can be overcome by using the Isolation Domain method to manage the LSA Type-5 flooding efficiency.
Junaidi Junaidi, Amirullah Andi Bramantya, Bintang Satya Pradipta
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 19; https://doi.org/10.12962/j24068535.v19i1.a962

Abstract:
Context – Internet of Things (IoT) interrelates computing devices, machines, animals, or people and things that use the power of internet usage to utilize data to be much more usable. Food is one of the mandatory human needs to survive, and most of it is produced by agriculture. Using IoT in agriculture needs appropriate software architecture that plays a prominent role in optimizing the gain. Objective and Method – Implementing a solution in a specific field requires a particular condition that belongs to it. The objectives of this research study are to classify the state of the art IoT solution in the software architecture domain perspective. We have used the Evidence- Based Software Engineering (EBSE) and have 24 selected existing studies related to software architecture and IoT solutions to map to the software architecture needed on IoT solutions in agriculture. Result and Implications – The results of this study are the classification of various IoT software architecture solutions in agriculture. The highlighted field, especially in the areas of cloud, big data, integration, and artificial intelligence/machine learning. We mapped the agriculture taxonomy classification with IoT software architecture. For future work, we recommend enhancing the classification and mapping field to the utilization of drones in agriculture since drones can reach a vast area that is very fit for fertilizing, spraying, or even capturing crop images with live cameras to identify leaf disease.
Syavira Tiara Zulkarnain, Nanik Suciati
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 19; https://doi.org/10.12962/j24068535.v19i1.a1031

Abstract:
Facial expression recognition (FER) on images with illumination variation and noises is a challenging problem in the computer vision field. We solve this using deep learning approaches that have been successfully applied in various fields, especially in uncontrolled input conditions. We apply a sequence of processes including face detection, normalization, augmentation, and texture representation, to develop FER based on Convolutional Neural Network (CNN). The combination of TanTriggs normalization technique and Adaptive Gaussian Transformation Method is used to reduce light variation. The number of images is augmented using a geometric augmentation technique to prevent overfitting due to lack of training data. We propose a representation of Modified Local Ternary Pattern (Modified LTP) texture image that is more discriminating and less sensitive to noise by combining the upper and lower parts of the original LTP using the logical AND operation followed by average calculation. The Modified LTP texture images are then used to train a CNN-based classification model. Experiments on the KDEF dataset show that the proposed approach provides a promising result with an accuracy of 81.15%.
Adenuar Purnomo, Handayani Tjandrasa
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 19; https://doi.org/10.12962/j24068535.v19i1.a1023

Abstract:
Deep learning is commonly used to solve problems such as biomedical problems and many other problems. The most common architecture used to solve those problems is Convolutional Neural Network (CNN) architecture. However, CNN may be prone to overfitting, and the convergence may be slow. One of the methods to overcome the overfitting is batch normalization (BN). BN is commonly used after the convolutional layer. In this study, we proposed a further usage of BN in CNN architecture. BN is not only used after the convolutional layer but also used after the fully connected layer. The proposed architecture is tested to detect types of seizures based on EEG signals. The data used are several sessions of recording signals from many patients. Each recording session produces a recorded EEG signal. EEG signal in each session is first passed through a bandpass filter. Then 26 relevant channels are taken, cut every 2 seconds to be labeled the type of epileptic seizure. The truncated signal is concatenated with the truncated signal from other sessions, divided into two datasets, a large dataset, and a small dataset. Each dataset has four types of seizures. Each dataset is equalized using the undersampling technique. Each dataset is then divided into test and train data to be tested using the proposed architecture. The results show the proposed architecture achieves 46.54% accuracy for the large dataset and 93.33% accuracy for the small dataset. In future studies, the batch normalization parameter will be further investigated to reduce overfitting.
Akbar Pandu Segara, Royyana Muslim Ijtihadie, Tohari Ahmad
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 19; https://doi.org/10.12962/j24068535.v19i1.a1011

Abstract:
Software Defined Network is a network architecture with a new paradigm which consists of a control plane that is placed separately from the data plane. All forms of computer network behavior are controlled by the control plane. Meanwhile the data plane consisting of a router or switch becomes a device for packet forwarding. With a centralized control plane model, SDN is very vulnerable to congestion because of the one-to-many communication model. There are several mechanisms for congestion control on SDNs, one of which is modifying packets by reducing the size of packets sent. But this is considered less effective because the time required will be longer because the number of packets sent is less. This requires that network administrators must be able to configure a network with certain routing protocols and algorithms. Johnson's algorithm is used in determining the route for packet forwarding, with the nature of the all-pair shortest path that can be applied to SDN to determine through which route the packet will be forwarded by comparing all nodes that are on the network. The results of the Johnson algorithm's latency and throughput with the comparison algorithm show good results and the comparison of the Johnson algorithm's trial results is still superior. The response time results of the Johnson algorithm when first performing a route search are faster than the conventional OSPF algorithm due to the characteristics of the all pair shortest path algorithm which determines the shortest route by comparing all pairs of nodes on the network.
Rizky Januar Akbar, Nurul Fajrin Ariyani, Adistya Azhar, Andika Andra
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 19; https://doi.org/10.12962/j24068535.v19i1.a1022

Abstract:
There is an impersonation (login as) feature in several applications that can be used by system administrators who have special privileges. This feature can be utilized by development and maintenance teams that have administrator rights to reproduce errors or bugs, to check specific features in applications according to the specific users’ login sessions. Beside its benefits, there is a security vulnerability that allows administrators to abuse the rights. They can access users’ private data or execute some activities inside the system without account or resource owners’ consents.This research proposes an impersonation method on authorization server using Client-Initiated Back-channel Authentication (CIBA) protocol. This method prevents impersonation without account or resource owners’ consent. The application will ask users’ authentication and permission via authentication device possessed by resource owners before the administrator performs impersonation. By utilizing authentication device, the impersonation feature should be preceded by users’ consent and there is no direct interaction needed between the administrator and resource owners to prove the users’ identities. The result shows that the implementation of CIBA protocol can be used to complement the impersonation method and can also run on the authorization server that uses OAuth 2.0 and OpenID Connect 1.0 protocols. The system testing is done by adopting FAPI CIBA conformance testing.
Nada Fitrieyatul Hikmah, Achmad Arifin, Tri Arief Sardjono
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a992

Abstract:
ECG signals have very features time-varying morphology, distinguished as P wave, QRS complex, and T wave. Delineation in ECG signal processing is an important step used to identify critical points that mark the interval and amplitude locations in the features of each wave morphology. The results of ECG signal delineation can be used by clinicians to associate the pattern of delineation point results with morphological classes, besides delineation also produces temporal parameter values of ECG signals. The delineation process includes detecting the onset and offset of QRS complex, P and T waves that represented as pulse width, and also the detection of the peak from each wave feature. The previous study had applied bandpass filters to reduce amplitude of P and T waves, then the signal was passed through non-linear transformations such as derivatives or square to enhance QRS complex. However, the spectrum bandwidth of QRS complex from different patients or same patient may be different, so the previous method was less effective for the morphological variations in ECG signals. This study developed delineation from the ECG feature extraction based on multiresolution analysis with discrete wavelet transform. The mother wavelet used was a quadratic spline function with compact support. Finally, determination of R, T, and P wave peaks were shown by zero crossing of the wavelet transform signals, while the onset and offset were generated from modulus maxima and modulus minima. Results show the proposed method was able to detect QRS complex with sensitivity of 97.05% and precision of 95.92%, T wave detection with sensitivity of 99.79% and precision of 96.46%, P wave detection with sensitivity of 56.69% and precision of 57.78%. The implementation in real time analysis of time-varying ECG morphology will be addressed in the future research.
Muhammad Ihsan Diputra, Ahmad Akbar Megantara, Pima Hani Safitri, Didik Purwanto
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a942

Abstract:
Wireless multimedia sensor network (WMSN) is one of broad wide application for developing a smart city. Each node in the WMSN has some primary components: sensor, microcontroller, wireless radio, and battery. The components of WMSN are used for sensing, computing, communicating between nodes, and flexibility of placement. However, the WMSN technology has some weakness, i.e. enormous power consumption when sending a media with a large size such as image, audio, and video files. Research had been conducted to reduce power consumption, such as file compression or power consumption management, in the process of sending data. We propose Green Communication (GeCom), which combines power control management and file compression methods to reduce the energy consumption. The power control management method controls data transmission. If the current data has high similarity with the previous one, then the data will not be sent. The compression method compresses massive data such as images before sending the data. We used the low energy image compression algorithm algorithm to compress the data for its ability to maintain the quality of images while producing a significant compression ratio. This method successfully reduced energy usage by 2% to 17% for each data.
Cosmas Haryawan, Maria Mediatrix Sebatubun
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a990

Abstract:
University is one of the educational institutions and can be established by the government or the individual. At this time, Indonesia has hundreds of universities spread throughout the region. As an educational institution, university of course must be able to educate its students and issue quality graduates with the academically and non-academically qualified. In its implementation, there are many problems that should be resolved as well as possible, such as when there are students who intentionally stop or disappear before completing their education or are even unable to complete their education and issued by institution (dropout).Based on these problems, this research makes a model for predicting students who have the potential to fail or dropout during their studies using one of the data mining methods namely Multilayer Perceptron by referring to personal and academic data. The results obtained from this research are 86.9% an accuracy rate with the 54.7% sensitivity, and 95.4% specificity. This research is expected to be used to determine the need strategies to minimize the number of students who stop or dropout.
Syukron Rifail Muttaqi, Bagus Jati Santoso
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a999

Abstract:
The increasing use of mobile devices makes spatial data worthy of consideration. To get maximum results, users often look for the best from a collection of objects. Among the algorithms that can be used is the skyline query. The algorithm looks for all objects that are not dominated by other objects in all of its attributes. However, data that has many attributes makes the query output a lot of objects so it is less useful for the user. k-dominant skyline queries can be a solution to reduce the output. Among the challenges is the use of skyline queries with spatial data and the many user preferences in finding the best object. This study proposes IKSR: the k-dominant skyline query algorithm that works in a road network environment and can process many queries that have the same subspace in one processing. This algorithm combines queries that operate on the same subspace and set of objects with different k values by computing from the smallest to the largest k. Optimization occurs when some data for larger k are precomputed when calculating the result for the smallest k so the Voronoi cell computing is not repeated. Testing is done by comparing with the naïve algorithm without precomputation. IKSR algorithm can speed up computing time two to three times compared to naïve algorithm.
Avin Maulana, Chastine Fatichah, Nanik Suciati
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a1004

Abstract:
Facial inpainting or face restoration is a process to reconstruct some missing region on face images such that the inpainting results still can be seen as a realistic and original image without any missing region, in such a way that the observer could not realize whether the inpainting result is a generated or original image. Some of previous researches have done inpainting using generative network, such as Generative Adversarial Network. However, some problems may arise when inpainting algorithm have been done on unaligned face. The inpainting result show spatial inconsistency between the reconstructed region and its adjacent pixel, and the algorithm fail to reconstruct some area of face. Therefore, an improvement method in facial inpainting based on deep-learning is proposed to reduce the effect of the stated problem before, using GAN with additional loss from feature reconstruction and two discriminators. Feature reconstruction loss is a loss obtained by using pretrained network VGG-Net, Evaluation of the result shows that additional loss from feature reconstruction loss and two type of discriminators may help to increase visual quality of inpainting result, with higher PSNR and SSIM than previous result.
Afrizal Laksita Akbar, Chastine Fatichah, Ahmad Saikhu
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a1000

Abstract:
Personal identification can be done by using face, fingerprint, palm prints, eye’s retina, or voice recognition which commonly called as biometric methods. Face recognition is the most popular and widely used among those biometric methods. However, there are some issues in the implementation of this method: lighting factor, facial expression, and attributes (chin, mustache, or wearing some accessories). In this study, we propose a combination method of Discrete Wavelet Transform and Stationary Wavelet Transform that able to improve the image quality, especially in the small-sized image. Moreover, we also use Histogram Equalization in order to correct noises such as over or under exposure, Discrete Cosine Transform in order to transform the image into frequency domain, and Deep Neural Networks in order to perform the feature extraction and classify the image. A 10-fold cross-validation method was used in this study. As the result, the proposed method showed the highest accuracy up to 92.73% compared to Histogram Equalization up to 80.73%, Discrete Wavelet Transform up to 85.85%, Stationary Wavelet Transform up to 64.27%, Discrete Cosine Transform up to 89.50%, the combination of Histogram Equalization, Discrete Wavelet Transform, and Stationary Wavelet Transform up to 69.77%, and the combination of Stationary Wavelet Transform, Discrete Wavelet Transform, and Histogram Equalization up to 77.39%.
Ari Mahardika Ahmad Nafis, Darlis Herumurti
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a989

Abstract:
Mathematic is basic but fundamental knowledge, but in fact many students do not have the motivation to learn it because they think mathematic is boring. Therefore, an innovation is needed to motivate students, one of them is by using an educational game. Racing, shooting and fighting games are the most popular types of games in 2019 according to InvisionCommunity. Shooting game is a genre that used a lot in the educational games for learning math, while racing game and fighting game are not used much for educational games. This research aims to develop and measure the effectiveness of the games from these three genre of games as a means of learning elementary arithmetic at the elementary school level. The effectiveness of an educational game can be observed from the increment in learning outcomes obtained after conducting an experiment. We can know the most effective type of game in this experiment by compare the improvement in learning outcomes after playing all three games. The comparative analysis will be carried out using ANOVA. In this research, we used data from 60 participant with elementary level of education between grade 1 to 3. The results were obtained by calculating the difference in the participants' initial scores obtained from before playing the game and participants’ final scores obtained after playing the educational game. The results show that educational racing games have the highest increase of 6.3% compared to shooter games with 3% increase or fighting games with increase of 4.3%.
Nafa Zulfa, Hafara Firdausi, Rakha Asyrofi
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a937

Abstract:
Supply chain management (SCM) system is an essential requirement for companies and manufacturers to collaborate in doing business. There are many techniques to manage supply chains, such as using Excel sheets and web-based applications. However, these techniques are ineffective, insecure, and prone to human error. In this paper, we propose CLOUTIDY, a cloud-based SCM system using SEMAR (Service Market) and Blockchain system. We modify JUGO architecture to develop SEMAR as a broker between users and cloud service providers. Also, we apply the Blockchain concept to store the activity log of the SCM system in a decentralized database. CLOUTIDY system can solve several common cases: service selection, resource provisioning, authentication and access control. Also, it improves the security of data by storing each activity log of the supply chain management system in the Blockchain system.
Dimas Ari Setyawan, Chastine Fatichah
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a1005

Abstract:
The classification process using a decision tree is a classification method that has a feature selection process. Decision tree classifications using information gain have a disadvantage when the dataset has unique attributes for each imbalanced class record and distribution. The data used for decision tree classification has 2 types, numerical and nominal. The numerical data type is carried out a discretization process so that it gets data intervals. Weaknesses in the information gain method can be reduced by using a dispersion ratio method that does not depend on the class distribution, but on the frequency distribution. Numeric type data will be dis-criticized using the hierarchical clustering method to obtain a balanced data cluster. The data used in this study were taken from the UCI machine learning repository, which has two types of numeric and nominal data. There are two stages in this research namely, first the numeric type data will be discretized using hierarchical clustering with 3 methods, namely single link, complete link, and average link. Second, the results of discretization will be merged again then the formation of trees with splitting attributes using dispersion ratio and evaluated with cross-validation k-fold 7. The results obtained show that the discretization of data with hierarchical clustering can increase predictions by 14.6% compared with data without discretization. The attribute splitting process with the dispersion ratio of the data resulting from the discretization of hierarchical clustering can increase the prediction by 6.51%.
Adhatus Solichah Ahmadiyah, Faridatul Aidah, Navinda Meutia, Denise Rahmadina, Daniel Lumbantobing, Ratih Anggraini
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a1007

Abstract:
Remarketing is a powerful tool for marketers to offer products over and over to existing customers or potential customers. By using remarketing, the marketers target to further down their sales funnel. As in traditional marketing, most online marketers find it challenging to deliver the best way of advertising their products according to what customers need or like. This research aims to achieve the right promotional media alternatives based on customer preferences. A clustering method was used to perform behavior segmentation on sales data. Then, customer reviews on the purchased products collected from online platforms were analyzed to obtain customer preferences. Finally, customer preference was mapped to some suitable promotion media. The experiment result showed that pipelining sales data and product reviews could obtain definite and distinct promotional media based on customer preference. Overall, this research may help online marketers bundle specific remarketing content into promotional media that matches to customer favorites.
Wahyu Suadi, Supeno Djanali, Waskitho Wibisono, Radityo Anggoro, Ary Mazharuddin Shiddiqi
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a986

Abstract:
State Drive (SSD) is an alternative to data storage that is popular today, widely used as a media cache to speed up data access to the hard disk (HDD). This paper proposes page replacement technique on SSD cache that used frequency and recency parameter, alternately. The algorithm is selected adaptively based on trace input. This method helps to overcome changes in access patterns while minimizing the number of write processes to SSD. The proposed algorithm can choose a replacement technique that suits the user access pattern so that it can bring a better hit rate. The proposed algorithm is also integrated with the ghost-cache mechanism so that the reduction in the number of writing processes to SSD is significant. The experiment runs using a real dataset, describing trace of data read, and data write taken from real usage. The trial shows that the proposed algorithm can give good results compared to other similar algorithms.
Muhammad Fattahilah Rangkuty, Royyana Muslim Ijtihadie, Tohari Ahmad
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a1008

Abstract:
SDN is a computer network approach that allows network administrators to manage network services through the abstraction of functionality at a higher level, by separating systems that make decisions about where traffic is sent (control plane), then forwarding traffic to the chosen destination (data plane). SDN can have problems with network congestion, high latency, and decreased throughput due to unbalanced traffic allocation on available links, so a load-balancing load method is needed. This technique divides the entire load evenly on each component of the network on the path or path that connects the data plane and S-D (Source Destination) host. The Least Loaded Path (LLP) of our proposed concept, which is a Dijkstra development, selects the best path by finding the shortest path and the smallest traffic load, the smallest traffic load (minimum cost) obtained from the sum of tx and rx data in the switchport data plane involved in the test, this result which will then be determined as the best path in the load balancing process.
Junaidi Junaidi, Prasetyo Wibowo, Dini Yuniasri, Putri Damayanti, Ary Mazharuddin Shiddiqi, Baskoro Adi Pratomo
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a940

Abstract:
A common way to maintain the quality of service on systems that are growing rapidly is by increasing server specifications or by adding servers. The utility of servers can be balanced with the presence of a load balancer to manage server loads. In this paper, we propose a machine learning algorithm that utilizes server resources CPU and memory to forecast the future of resources server loads. We identify the timespan of forecasting should be long enough to avoid dispatcher's lack of information server distribution at runtime. Additionally, server profile pulling, forecasting server resources, and dispatching should be asynchronous with the request listener of the load balancer to minimize response delay. For production use, we recommend that the load balancer should have friendly user interface to make it easier to be configured, such as adding resources of servers as parameter criteria. We also recommended from beginning to start to save the log data server resources because the more data to process, the more accurate prediction of server load will be.
Muh. Anas Faishol, Endroyono Endroyono, Astria Nur Irfansyah
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a988

Abstract:
Air is one of the primary needs of living things. If the condition of air is polluted, then the lives of humans and other living things will be disrupted. So it is needed to perform special handling to maintain air quality. One way to facilitate the prevention of air pollution is to make air pollutionforecasting by utilizing past data. Through the Environmental Office, the Surabaya City Government has monitored air quality in Surabaya every 30 minutes for various air quality parameters including CO, NO, NO2, NOx, PM10, SO2 and meteorological data such as wind direction, wind direction, wind speed, wind speed, global radiation, humidity, and air temperature. These data are very useful to build a prediction model for the forecast of air pollution in the future. With the large amount and variance of data generated from monitoring air quality in Surabaya city, a qualified algorithm is needed to process it. One algorithm that can be used is Recurrent Neural Network - Long Short Term Memory (RNN-LSTM). RNN-LSTM is built for sequential data processing such as time-series data. In this study, several analyses are performed. There are trend analysis, correlation analysis of pollutant values to meteorological data, and predictions of carbon monoxide pollutants using the Recurrent Neural Network - LSTM in the city of Surabaya correlated with meteorological data. The results of this study indicate that the best prediction model using RNN-LSTM with RMSE calculation gets an error of 1,880 with the number of hidden layer 2 and epoch 50 scenarios. The predicted results built can be used as a reference in determining the policy of the city government to deal with air pollution going forward.
Bahri Rizaldi, Doni Setio Pambudi, Taufiqotul Bariyah
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i2.a897

Abstract:
Currently, route search is made easier by the presence of a Global Positioning System (GPS) technology that can be used by using the Maps application on a smartphone. By using the Maps application, people can find out their current location and can find a route to their desired destination. But the level of GPS accuracy will decrease if the user is in a building or in a closed room. This is caused by the satellite signals being sent that are not able to penetrate thick walls or concrete so that the search for routes using GPS is limited to the search for routes outside the building or outdoors. In this research, Bluetooth Low Energy and trilateration are used to determine the location in a room or building and Dijkstra's algorithm for finding the shortest route to the destination location. The proposed method has a location determination error of 0.728 meters with a distance between the user and the beacon less than 10 meters to get a stable signal.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i1.a930

Abstract:
Ketersediaan bandwidth merupakan salah satu aspek penting untuk menjamin QoS dalam transmisi data, terutama pada jaringan nirkabel. Walaupun demikian, prediksi ketersediaan pada jaringan nirkabel masih sulit dilakukan karena medium transmisi dapat digunakan oleh beberapa node secara bersamaan. Selain itu jaringan nirkabel juga rentan terhadap pengaruh dari sinyal transmisi yang dihasilkan dari node lain, terutama hidden node. Beberapa penelitian telah dilakukan untuk mengembangkan model prediksi ketersediaan bandwidth. Walaupun demikian, belum terdapat mekanisme terstandarisasi yang digunakan untuk mengevaluasi ketersediaan bandwidth pada jaringan nirkabel. Selain itu tingkat akurasi dari setiap model juga masih belum diketahui ketika diimplementasikan pada jaringan nirkabel, terutama dengan keberadaan hidden node. Oleh karena itu penelitian ini berupaya untuk menginvestigasi kinerja dari setiap model untuk memprediksi ketersediaan bandwidth pada jaringan nirkabel dengan interferensi hidden node. Model prediksi yang dibandingkan adalah Distributed Lagrange Interpolation Based Available Bandwidth Estimation (DLI-ABE), Cognitive Passive Estimation Of The Available Bandwidth (cPEAB), Improved Available Bandwidth (IAB), dan Available Bandwidth Estimation (ABE). Percobaan dilakukan dalam skala simulasi yang dikembangkan menggunakan simulasi jaringan OMNet++. Hasil penelitian menunjukan bahwa model ABE memperoleh tingkat akurasi yang paling baik sebesar 85,25%.
Reza Prasetya Prayogo, Joko Lianto Buliali
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i1.a902

Abstract:
Terdapat beberapa titik acuan dalam satu rute penerbangan untuk keperluan navigasi yang disebut waypoint. Pada penelitian ini penulis melakukan segmentasi untuk membagi satu rute penerbangan (Surabaya-Palu) menjadi 7 segmen yang terdiri dari 8 waypoint, dengan membuat garis imajiner secara tegak lurus melewati masing-masing waypoint. Pada tiap segmen dilakukan analisa terkait lokasi yang paling sering dilalui menggunakan pendekatan clustering.Dalam penelitian ini penulis menggunakan algoritma clustering K-means dengan optimasi centroid yang mengimplementasikan algoritma Ant Lion Optimizer (ALO) atau disebut dengan K-means-ALO. Jumlah cluster ditentukan sebelumnya, kemudian dilakukan validasi pengelompokan internal dengan menggunakan silhouette index. Hasil metode pengelompokan diuji nilai performansinya. Hasil akhir dari jumlah cluster yang sudah ditentukan diambil nilai validitas cluster terbaik yaitu jumlah cluster yang optimum pada tiap segmen area penerbangan.Pengujian dilakukan dengan membandingkan nilai silhouette index untuk setiap percobaan jumlah cluster terhadap kedua metode yaitu K-means dan K-means-ALO. Pada uji coba yang dilakukan, metode optimasi yang diusulkan menghasilkan validitas cluster yang lebih baik sesuai nilai silhouette index pada tiga segmen, yaitu segmen 2, 3, dan 5 akan tetapi signifikan di semua segmen berdasarkan uji statistik Analysis of Variance (ANOVA) dan uji lanjut Least Significant Difference (LSD).
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i1.a944

Abstract:
Greedy Perimeter Stateless Routing (GPSR) merupakan protokol routing yang memiliki performa baik di lingkungan VANET. Protokol GPSR memiliki kelemahan ketika node selanjutnya mengalami keadaan local maximum, yaitu ketika node selanjutnya tidak dapat mengirim paket ke node selanjutnya dikarenakan tidak ada node di sekitar yang memiliki posisi terdekat dengan node tujuan atau semua node terdekat sudah pernah menerima paket tersebut. Untuk mengatasi keadaan local maximum, protokol GPSR dimodifikasi dengan konsep overlay network. Overlay network diterapkan dengan menggunakan route discovery milik protokol dynamic source routing (DSR). Hasil dari proses route discovery akan menjadi acuan pencarian virtual anchor point (VAP). VAP merupakan representasi dari overlay network dan berguna untuk mengganti acuan posisi node tujuan dalam metode greedy forwarding. Selain VAP, pemilihan node selanjutnya menerapkan metode area optimum dalam pemilihan node selanjutnya untuk menghindari node yang berada pada luar batas transmisi node pengirim. Dalam makalah ini, evaluasi dilakukan terhadap kinerja routing protocol pada skenario real. Hasil simulasi diukur berdasarkan nilai rata-rata packet delivery rasio (PDR), end to end delay dan routing overhead (RO). Jumlah node yang digunakan dalam simulasi dimulai dari 50, 100, dan 150 node. Dari simulasi yang dilakukan didapatkan bahwa real performa GPSR modifikasi dibandingkan dengan GPSR tradisional mengalami peningkatan nilai PDR sebesar 72%, tetapi terjadi peningkatan pada nilai rata-rata end to end delay sebesar 1118% dan peningkatan nilai rata-rata RO sebesar 0.6%.
, Radityo Anggoro
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i1.a976

Abstract:
VANETs (Vehicular ad hoc networks) merupakan teknologi yang digunakan untuk komunikasi antar kendaraan. Dalam VANETs, kendaraan bergerak dengan kecepatan tinggi dan arah yang berbeda-beda, sehingga menyebabkan jaringan komunikasi yang telah terbentuk berubah dengan cepat. Salah satu contoh protokol routing yang sering digunakan dalam komunikasi VANETs adalah GPSR. Greedy perimeter stateless routing (GPSR), routing protokol baru untuk jaringan nirkabel yang menggunakan informasi posisi node tujuan untuk membuat keputusan penerusan paket. Topologi jaringan VANETs yang sangat dinamis menyebabkan komunikasi antar node tidak stabil. Node next hop yang telah dipilih mungkin telah keluar dari jangkauan transmisi sebelum menerima paket. Penelitian ini bertujuan untuk memecahkan masalah dalam protokol routing GPSR terkait ketidakstabilan komunikasi karena posisi node yang berubah dalam mode greedy forwarding. Dalam penelitian ini penulis menambahkan set parameter yang digunakan untuk mengambil keputusan routing dengan memasukkan faktor kecepatan dan arah pergerakan kendaraan. Setiap node akan melakukan penghitungan rata-rata geometrik kecepatannya sebelum penyiaran beacon message. Informasi rata-rata geometrik tersebut akan ditambahkan pada beacon message. Setelah node menerima paket beacon message, node akan memperbarui informasi yang terdapat pada neighbor table. Pada penelitian ini, neighbor table juga menyimpan informasi posisi node pada waktu sebelumnya dan informasi rata-rata geometrik kecepatan. Informasi dalam neighbor table tersebut akan digunakan dalam pengambilan keputusan routing. Penelitian ini juga melakukan modifikasi pada metode penerusan paket greedy forwarding. Jika penerusan paket menggunakan metode greedy forwarding, tahapan pertama yang dilakukan adalah menemukan node-node yang berada dalam area komunikasi optimum. Selanjutnya dari node-node yang berada pada area komunikasi optimum, akan dilakukan seleksi pemilihan forwarding node. Node yang layak menjadi forwarding node adalah node yang paling minimal perbedaan rata-rata geometrik kecepatannya dan bergerak mendekati node tujuan. Diharapkan dengan penelitian ini dapat meningkatkan performa protokol routing GPSR
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i1.a841

Abstract:
Leaks in water pipeline networks have cost billions of dollars each year. Robust leak quantification (to detect and to localize) methods are needed to minimize the lost. We quantify leaks by classifying their locations using machine learning algorithms, namely Support Vector Machine and C4.5. The algorithms are chosen due to their high performance in classification. We simulate leaks at different positions at different sizes and use the data to train the algorithms. We tune the algorithm by optimizing the algorithms' parameters in the training process. Then, we tested the algorithms' models against real observation data. We also experimented with noisy data, due to sensor inaccuracies, that often happen in real situations. Lastly, we compared the two algorithms to investigate how accurate and robust they localize leaks with noisy data. We found that C4.5 is more robust against noisy data than SVM.
Yuni Tjung, Ardianto Wibowo,
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 18; https://doi.org/10.12962/j24068535.v18i1.a898

Abstract:
Data utilization has effectively contributed for institutions growth by providing insights for managerial purposes. In Information Technology (IT) Department of Politeknik Caltex Riau, information systems were built separately, makes it hard for the head of study program to analyze academic performance. For analytical purposes, there’s a business intelligence developed to equip each head of study programs in IT Department with knowledge about their department. Unfortunately, the business intelligence hasn’t considered with data integration. To solve this problem, this research proposes 2 different academic near-real time data integration model that are documented using Enterprise Integration Pattern and benchmarkes the implementation to obtain best data integration model. The models use Message-Oriented Middleware, a technology that enables asynchronous communication between diverse applications. This research uses WSO2 ESB as the MOM tools in Service-Oriented Architecture (SOA) that use NuSOAP library for helping generating web service WSDL and will use Enterprise Application Integration approach. The testing is conducted based on ISO 9126 aspects: functionality, efficiency, and reliability. Based on the testing results, it can be concluded that both integration models fulfill the functionality and reliability aspects, but the 2nd pattern is more efficient because it distincts message channel and store for each dimension and fact table.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i2.a823

Abstract:
Aspek kualitas kebutuhan non-fungsional merupakan salah satu faktor penting yang berperan dalam kesuksesan pengembangan perangkat lunak. Namun, mengidentifikasi aspek kualitas kebutuhan non-fungsional merupakan hal yang sulit untuk dilakukan. Karena aspek kualitas kebutuhan non-fungsional sering ditemukan tercampur dengan kebutuhan fungsional. Oleh karena itu dibutuhkan suatu cara untuk dapat mengidentifikasi aspek kualitas kebutuhan non-fungsional. Penelitian yang ada mampu mengidentifikasi aspek kebutuhan non-fungsional dengan melakukan klasifikasi. Akan tetapi, standar kualitas yang digunakan sebagai rujukan untuk melabeli kalimat kebutuhan masih menggunakan standar ISO/IEC 9126. ISO/IEC 9126 merupakan standar lama yang dirilis pada tahun 2001. Peneliti sebelumnya mengungkapkan ambiguitas dalam enam sub-atribut pada struktur hirarkis ISO/IEC 9126. Oleh karena itu, standar kualitas yang digunakan untuk melabeli kalimat kebutuhan pada penelitian ini adalah ISO/IEC 25010. Sedangkan metode klasifikasi yang digunakan adalah FSKNN. Metode klasifikasi yang digunakan diuji dengan menggunakan nilai tetangga terdekat 10, 20 dan 30. Pada penelitian ini metode FSKNN berhasil memeroleh nilai tertinggi berdasarkan ground truth pakar yaitu precision sebesar 22.55 dan recall 27.64.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i2.a788

Abstract:
Perpustakaan ITS adalah salah satu penyedia jasa informasi di ITS.  Berbagai koleksi fisik yang dikelola meliputi buku teks, buku tugas akhir, buku tesis, jurnal, majalah, serta prosiding seminar nasional. Setiap tahunnya, perpustakaan ITS memperoleh alokasi dana untuk  pengadaan buku cetak sebesar 1 M, e-journal sebesar 6 M, dan 300 juta untuk pengadaan e-book. Akan tetapi, dana tidak terserap dengan baik dan feedback untuk pengadaan bahan pustaka ke ULP tidak berjalan maksimal dikarenakan pustakawan mengalami kesulitan ketika melakukan proses seleksi judul-judul bahan pustaka yang akan diajukan ke ULP untuk dibeli. Hal ini menyebabkan bahan pustaka, khususnya buku, yang dibeli kebanyakan tidak sesuai dengan kebutuhan pengguna. Untuk itu diperlukan upaya mencari informasi buku baru sebagai bahan pustaka yang sesuai dengan kebutuhan pengguna berbasis teknologi informasi. Berdasarkan data pengadaan buku di perpustakaan ITS lebih didominasi oleh buku pengembangan yang mendukung referensi publikasi ilmiah. Publikasi ilmiah yang dilakukan oleh para dosen mayoritas merupakan luaran dari penelitian dosen. Oleh karena itu, pada penelitian ini diusulkan klasterisasi tren topik penelitian sebagai rekomendasi pengadaan bahan pustaka di Perpustakaan ITS. Penelitian ini menerapkan konsep text mining yang terdiri dari beberapa tahapan proses yaitu: text preprocessing, proses ekstraksi fitur, proses clustering, dan post-processing. Text preprocessing dilakukan untuk memperbaiki kualitas data teks, sehingga dapat menghasilkan klaster yang relevan dan akurat. Langkah-langkah pada tahap text preprocessing adalah case folding, tokenizing, filtering, dan stemming. Kemudian, dilakukan proses ekstraksi fitur yaitu dengan teknik pembobotan menggunakan Term Frequency dan Inverse Document Frequency (TF-IDF). Fitur-fitur yang dihasilkan pada tahap ekstraksi fitur dilakukan proses clustering menggunakan metode unsupervised learning untuk menghasilkan klaster topik penelitian. Tahap post-processing dilakukan untuk mengevaluasi dan menganalisa hasil klasterisasi tersebut yang selanjutnya digunakan sebagai rekomendasi pengadaan bahan pustaka, khususnya buku.
Yandhy Raka Mulyawan, Caecilia Citra Lestari
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i2.a791

Abstract:
Banyak ibu rumah tangga yang kebingungan untuk menentukan masakan apa yang akan mereka masak sehingga bahan makanan yang mereka miliki menjadi rusak akibat tidak kunjung dimasak. Sebagian besar ibu rumah tangga mendapatkan ide resep dari website resep karena mudah untuk diakses dan memiliki resep yang cukup lengkap, namun kelemahannya kebanyakan dari website resep tidak memiliki fitur untuk pencarian resep berdasarkan bahan-bahan yang dimiliki. Aplikasi telepon genggam dipilih untuk memecahkan masalah tersebut. Pada penelitian ini penulis akan membuat rancang bangun sistem rekomendasi resep masakan berdasarkan bahan baku dengan menggunakan algoritma penyaringan berbasis konten (CBFA). Algoritma ini merekomendasikan resep yang memiliki kesamaan dengan bahan makanan yang dimasukkan oleh pengguna. Aplikasi dibuat menggunakan file PHP untuk memproses data resep, seperti query data, mengecek data yang sama, menentukan weight serta menghitung dan mengurutkan resep menurut CBFA. Hasil dari pengujian menunjukkan bahwa rekomendasi resep sudah sesuai dengan kekuatan 71%.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i2.a825

Abstract:
This study discusses an application of fuzzy logic in educational game. This game requires a fast player response. We made this game in order to raise children awareness to encounter strangers. In this game, strangers are enemies in the form of hand drawings. The main characters in this game are boy and girl. The player can choose a character as desired. Each selected character has a different level. At each level, there is a level of difficulty based on the ease of passing enemies to reach the goal. The player would win the game once he/she manages to escape from strangers and arrives at home. Each level of the game has an enemy who tries to approach the player. If the enemy caught by the player, it will get consequences according to the fuzzy rules that applied to the game. The application of fuzzy logic in this game is to regulate the form of reward that will receive by the player. The basis for determining reward is the living conditions, time, and scores obtained by the player. In this research, we use fuzzy Sugeno logic for giving rewards. We conclude that fuzzy logic applies to our educational game.
Muhammad Riduwan, ,
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i2.a892

Abstract:
Jumlah penelitian di dunia mengalami perkembangan yang pesat, setiap tahun berbagai peneliti dari penjuru dunia menghasilkan karya ilmiah seperti makalah, jurnal, buku dsb. Metode klasterisasi dapat digunakan untuk mengelompokkan dokumen karya ilmiah ke dalam suatu kelompok tertentu berdasarkan relevansi antar topik. Klasterisasi pada dokumen memiliki karakteristik yang berbeda karena tingkat kemiripan antar dokumen dipengaruhi oleh kata-kata pembentuknya. Beberapa metode klasterisasi kurang memperhatikan nilai semantik dari kata. Sehingga klaster yang terbentuk kurang merepresentasikan isi topik dokumen. Klasterisasi dokumen teks masih memiliki kemungkinan adanya outlier karena pemilihan fitur teks yang tidak optimal. Oleh karena itu dibutuhkan pemrosesan data yang tepat serta metode yang mengoptimalkan hasil klaster. Penelitian ini mengusulkan metode klasterisasi dokumen menggunakan Weighted K-Means yang dipadukan dengan Maximum Common Subgraph. Weighted k-means digunakan untuk klasterisasi awal dokumen berdasarkan kata-kata yang diekstraksi. Pembentukan Weighted K-Means berdasarkan perhitungan Word2Vec dan TextRank dari kata-kata dalam dokumen. Maximum common subgraph merupakan tahap pembentukan graf yang digunakan dalam penggabungan klaster untuk menghasilkan klaster baru yang lebih optimal. pembentukan graf dilakukan dengan perhitungan nilai Word2vec dan Co-occurrence dari klaster. Representasi topik dokumen tiap klaster dapat dihasilkan dari pemodelan topik Latent Dirichlet Allocation (LDA). Pengujian dilakukan dengan menggunakan dataset artikel ilmiah dari Scopus. Hasil dari analisis Koherensi topik menunjukkan nilai koherensi usulan metode adalah 0.532 pada dataset 1 yang bersifat homogen dan 0.472 pada dataset 2 yang bersifat heterogen.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i2.a857

Abstract:
Diabetic Retinopathi adalah kelainan pembuluh darah retina pada mata yang diakibatkan komplikasi penyakit diabetes. Deteksi lebih dini diperlukan agar kelainan ini dapat ditangani secara cepat dan tepat. Kelainan ini ditandai dengan melemahnya bagian pembuluh darah tipis akibat tersumbatnya aliran darah kemudian menyebabkan bengkak pada mata bahkan kebutaan. Oleh karena itu diperlukan metode analisa pembuluh darah retina melalui proses segmentasi pembuluh darah terutama pada bagian penting yaitu pembuluh darah tipis. Peneliti mengusulkan penggabungan metode perbaikan pembuluh darah tipis atau yang dikenal dengan Thin Vessel Enhancement dan Fuzzy Entropy. Thin Vessel Enhancement berfungsi untuk memperbaiki citra agar dapat mengekstrak lebih banyak bagian pembuluh darah khususnya pembluh darah tipis, sedangkan Fuzzy Entropy dapat menentukan nilai optimal threshold berdasarkan nilai entropy pada membership function. Segmentasi yang dihasilkan dibagi menjadi 3 kategori yaitu pembuluh darah utama, medium, dan tipis. Uji coba dilakukan terhadap metode Thin Vessel Enhancement menggunakan 1 kernel dan Fuzzy Entropy dari nilai threshold ke-1 maka diperoleh nilai accuracy, sensitivity, dan specivicity sebesar 94.81%, 66.83%, dan 97.51%.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a731

Abstract:
Sistem penggajian umumnya merupakan proses penggajian dari awal sampai gaji itu dibayarkan sesuai dengan hak masing-masing karyawan dapat dilakukan atau dilaksanakan dengan sistematis dan diharapkan tidak akan terdapat kesalahan dalam prosedur penggajian. Pada PT. Bumi Sawindo Permai (BSP) sebagai perusahaan pengolahan kelapa sawit yang menjadi tempat penelitian memiliki 3 unit bisnis yang ditunjang oleh pemanfaatan teknologi informasi, meliputi: perangkat lunak Enterprise Resource Planning ASCEND, dan Standart Operational Procedure (SOP) ASCEND, serta SOP Sistem Menejemen Integrasi (SMI). Proses bisnis produksi dan operasional tidak didefinisikan pada semua komponen tersebut. BSP memberlakukan sistem penggajian secara sistematis sesuai dengan prosedur yang berlaku, akan tetapi prosedur ini tidak tercantum dengan jelas kedalam sebuah blueprint. Ditemukan ketidakselarasan dari ketiga komponen tersebut pada implementasinya. Hal ini mereduksi indeks kinerja perusahaan khususnya pada proses penggajian karyawan yang dilakukan oleh bagian Human Capital Management (HCM) di 3 unit bisnis BSP. Dengan demikian perlu dilakukan eliminasi bottleneck pada proses bisnis berjalan. Pemodelan proses bisnis dengan Business Process Modelling Notation (BPMN) dilakukan sebagai langkah strategis pengembangan SOP sebagai standar baku mutu untuk peningkatan efisiensi perusahaan dalam hal pengelolaan sistem pembayaran gaji karyawan yang lebih terukur.
Dian Septiani Santoso, R.V. Hari Ginardi
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a785

Abstract:
Focused Web Crawler merupakan metode pencarian website yang sesuai dengan pencarian yang diinginkan oleh user. Untuk mendapatkan kecocokan yang baik, waktu yang dibutuhkan metode Focused Web Crawler lebih lama dibandingkan dengan metode pencarian web crawler pada umumnya yaitu algoritma Depth First Search (DFS) maupun Breadth First Search (BFS). Untuk mengatasi hal tersebut maka muncul sebuah ide yakni teknik pencarian Focused Web Crawler dengan menggunakan metode metaheuristic pencarian cuckoo yang dipadukan dengan pencarian pada data history pencarian yang disimpan. Namun dengan adanya penyimpanan data pada setiap kali pencarian link maka data akan semakin bertambah. Oleh karena itu diperlukan sebuah cara untuk mengurangi kebutuhan ruang penyimpanan. Cara yang dilakukan untuk mengurangi ruang penyimpanan dan tidak mengurangi nilai informasi dari data penyimpanan sebelumnya adalah dengan melakukan kompresi data. Dalam penelitian ini diusulkan metode kompresi data dengan melakukan kompresi multilevel menggunakan dua metode kompresi yaitu pengurangan prefix dan postfix kata dan kompresi string berbasis kamus dengan melakukan pembuatan indeks kamus kata. Hasil kompresi string kamus kata berupa data encode. Untuk menguji hasil dari kompresi data yaitu dengan melakukan perbandingan hasil pencarian link menggunakan metode Knutt Morris Pratt (KMP) dari data yang belum terkompresi dengan data yang telah terkompresi. Hasilnya didapatkan bahwa maksimum presisi dengan nilai 1 dan recall sebesar 0,73. Dari hasil percobaan metode, didapatkan bahwa rasio kompresi file rata-rata adalah sebesar 36,4%.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a820

Abstract:
Evaluasi kualitas fungsional dan antar muka situs web e-commerce dari perspektif pengguna sangat penting untuk membangun atau mengembangkan situs web e-commerce yang memenuhi standar kualitas. Namun, untuk menilai kualitas fungsional dan antar muka dari situs web e-commerce sulit untuk didefinisikan sehingga membutuhkan model evaluasi perangkat lunak. Pentingnya evaluasi kualitas situs web e-commerce berdasarkan karakteristik perangkat lunak utnuk dapat dikembangkan dan menyesuaikan standar kualitas perangkat lunak.Penelitian ini mengusulkan sebuah model evaluasi kualitas situs web e-commerce berdasarkan karakteristik pada functional suitability, performance efficiency, reliability dan usability pada ISO/IEC 25010. Pada penelitian ini menggunakan metode Fuzzy Mamdani untuk menilai kualitas dari situs web e-commerce berdasarkan karakteristik dan pembobotan kepentingan karakteristik menggunakan metode Analytical Hierarchy Process. Model yang diusulkan diterapkan ke beberapa situs web e-commerce di Indonesia sebagai studi kasus untuk mengevaluasi tingkat kualitas perangkat lunak. Hasil yang didapat dari model evaluasi dapat membantu pengembang untuk merancang dan menggembangkan situs web e-commerce yang kualitas dengan tingkat accurasy 0,684%.
Margaretha - Mei Mei,
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a734

Abstract:
Event organizer is a company which engaged in event organizing, PR, and advertising. A suitable information systems that fit company's business needs are required so the company can adapt themselves in this globalization era. Information systems had an important role in order to support the company’s business process and its performance. An example of applied information systems that have been widely used by companies is enterprise architecture. Currently, enterprise architecture has been used by many enterprises to be able connect between the planning and the technology implementation to the current business in the company. Enterprise Architecture Planning (EAP) is a method in enterprise architecture which can provides alignment between business and information technology by defining company’s needs. Main components of EAP are data architecture, applications, and technology. The framework that will be used to develop EAP on this research is TOGAF ADM with phases starting from preliminary phase, requirements management, architecture vision, business architecture, information systems architectures, technology architecture, opportunities & solutions, and migration planning. A result on this research will be a blueprint of enterprise architecture model that an event organizer can use in supporting its business. The blueprint contains the current used information systems and the ideal information systems planned by the authors.
Yuna Sugianela,
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a819

Abstract:
Buku-buku kuno Bahasa Jawa memiliki konten kekayaan intelektual Indonesia seperti agama, linguistik, filosofi, mitos, pelajaran moral, hukum dan norma adat, kerajaan, cerita rakyat, sejarah, dan lain sebagainya. Tidak banyak yang mempelajari karya tersebut karena ditulis dengan Aksara Jawa dan tidak banyak yang memahami. Untuk membantu penerjemahan dokumen berbahasa Jawa dilakukan otomatisasi sistem penerjemahan. tahap penerjemahan terdiri dari segmentasi untuk mendapatkan karakter dari citra tulisan dalam naskah Aksara Jawa. Kemudian tiap karakater dikenali sebagai abjad. Dan yang terakhir adalah mengkombinasikan tulisan latin yang telah dikenali menjadi kata yang berarti. Penelitian yang membahas tentang pengenalan Aksara Jawa telah dilakukan, seperti fokus pada segmentasi karakter dan pengenalan Aksara Jawa. Pada penelitian sebelumnya dilakukan perbaikan pada metode segmentasi namun tetap mendapatkan hasil yang sama dalam hal akurasi kebenaran. Pada penelitian kali ini diusulkan metode baru pada tahap ekstraksi fitur, yaitu menggunakan metode Histogram of Oriented Gradient (HOG). Metode HOG banyak digunakan pada pengenalan wajah, hewan, dan deteksi citra kendaraan, dan lain-lain. Penelitian ini juga pernah diusulkan untuk mengenali tulisan tangan berbahasa Inggris dan Huruf Bengali dan mendapatkan hasil yang optimal. Pada penelitian ini didapatkan hasil akurasi pengenalan karakter Aksara Jawa sebesar 89,7%.Ekstraksi Fitur, Histogram of Oriented Gradient, Aksara Jawa
, Tosca Yoel Connery
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a824

Abstract:
Recently, the development of the query-based preferences has received considerable attention from researchers and data users. One of the most popular preference-based queries is the skyline query, which will give a subset of superior records that are not dominated by any other records. As the developed version of skyline queries, a reverse skyline query rise. This query aims to get information about the query points that make a data or record as the part of result of their skyline query. Furthermore, data-oriented IT development requires scientists to be able to process data in all conditions. In the real world, there exist incomplete multidimensional data, both because of damage, loss, and privacy. In order to increase the usability over a data set, this study will discuss one of the problems in processing reverse skyline queries over incomplete data, namely the "why-not" problem. The considered solution to this "why-not" problem is advice and steps so that a query point that does not initially consider an incomplete data, as a result, can later make the record or incomplete data as part of the results. In this study, there will be further discussion about the dominance relationship between incomplete data along with the solution of the problem. Moreover, some performance evaluations are conducted to measure the level of efficiency and effectiveness.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a775

Abstract:
Setting Examination schedules to support learning evaluation is crucial. The ideal scheduling for this exam must be able to allocate all related components in the implementation of the test within a predetermined time span. The components of the implementation of an examination in a university include the departments in the faculty, a number of courses and participants, the room used, the time of execution, and the lecturer serving as supervisor. The arrangement of each component of the implementation of the exam needs to be carried out appropriately so there is no collision of the schedule between the participants, the schedule, the room used, and the supervisor in charge. The purpose of this study is to produce an ideal exam scheduling and examination supervisor. The study was conducted by applying the Multi-Criteria Iterative Forward Search from the Academic Information System (SIAKAD) data at the Faculty of Science and Technology, Unisnu Jepara. This research has resulted in a system that is able to create an examination schedule and supervisory schedule that accommodates all factors without conflict, well tested, and applied.
, Heru Sokoco, Yani Nurhadryani
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a717

Abstract:
Teknologi informasi (IT) dapat memperkuat organisasi. Termasuk Institut Pertanian Bogor (IPB) yang merupakan organisasi pendidikan tinggi. IPB telah memiliki sistem informasi yang terdiri dari 36 aplikasi. Infrastruktur yang memadai merupakan salah satu kunci kesuksesan untuk mendapatkan kinerja dan pelayanan TI yang baik. kebutuhan infrastruktur dapat dipenuhi dengan melakukan evaluasi terhadap tatakelola TI. Penelitian ini melakukan evaluasi tatakelola TI dengan mengukur tingkat kematangan tatakelola TI, dengan menggunakan framework COBIT 5 dan tingkat kepuasan pengguna layanan TI, dengan menggunakan serqual model. Tingkat kematangan TI menggunakan 13 proses pada COBIT 5 yaitu: EDM04,APO01, APO07, APO12, APO13, BAI04, BAI06, BAI09, BAI10, DSS01, DSS03, DSS05, dan MEA01. Tingkat kepuasan pengguna dilihat dari 4 dimensi yaitu: tangibles, reliability, responsiveness, dan assurance. Responden yang pada pengukuran kepuasan layanan TI adalah mahasiswa dan dosen sebanyak 100 responden. Hasil penelitian ini didapatkan tingkat kematangan tatakelola TI 2 proses berada pada level 0, 8 proses berada pada level 1, dan 3 proses berada pada level 2. Tingkat kepuasan pengguna terhadap layanan TI berada dibawah nilai harapan. Hasil tersebut dianalisis dengan matrik SWOT untuk melihat kekuatan, kelemahan, peluang dan ancaman. Kemudian disusun rekomendasi dengan mengacu kepada ITIL V3 2011
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 17; https://doi.org/10.12962/j24068535.v17i1.a771

Abstract:
Requirements engineering phase in software development resulting in a SRS (Software Requirements Specification) document.The use of natural language approach in generating such document has some drawbacks that caused 7 common mistakes among the engineer which had been formulated by Meyer as "The 7 sins of specifier". One of the 7 common mistakes is noise. This study attempted to detect noise in software requirements with spectral clustering. The clustering algorithm working on fewer dimensions compared to others. The resulting kappa coefficient is 0.4426.The result showed that the consistency between noise prediction and noise assessment made by three annotators is still low.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a688

Abstract:
Pengawasan terhadap kinerja akademik mahasiswa sebagai bentuk peningkatan mutu harus dilakukan secara tersistem dan terintegrasi. Namun pengawasan tersebut akan lebih efektif jika dilakukan secara periodik, yaitu pada tahun kedua dan keeempat dengan maksud pihak akademik dapat mengetahui perkembangan pencapaian belajar masing-masing mahasiswa dengan cepat, sehingga peringatan atau tindakan yang akan diberikan pada mahasiswa dan evaluasi terhadap seluruh aktivitas akademis dapat segera dilakukan. Penelitian ini bertujuan mengklasifikasikan kinerja mahasiswa melalui IPK dan jumlah SKS yang belum diselesaikan selama masa studinya. Klasifikasi dilakukan pada tahun kedua dan tahun keempat masa studi. Pada tahun kedua, klasifikasi dibagi menjadi tiga status yaitu Normal, Bermasalah dan Peringatan dengan ditentukan melalui standar nilai yang ditentukan. Pada tahun keempat klasifikasi dibagi menjadi dua kelompok, yaitu kelompok Lulus dan Drop Out. Proses klasifikasi yang dilakukan pada tahun keempat, menggunakan algoritma Naïve Bayes yang terbukti memiliki tingkat keakurasian yang tinggi dengan metode maximum likehood atau berdasarkan kemiripan tertinggi dari data yang diolah. Pada proses penghitungan Naïve Bayes terdapat sedikit keraguan apabila ada peluang yang bernilai 0. Oleh karena itu untuk memaksimalkan performa dari Naïve Bayes dalam klasifikasi ini, maka digunakan Dirichlet Smoothing. Berdasarkan hasil pengujian terhadap 200 data uji maka didapatkan nilai akurasi mencapai 91.50%, nilai precision sebesar 88.78% dan nilai recall adalah 95%. Dengan demikian dapat diketahui bahwa data memiliki nilai yang konsisten.
Mirza Ramadhani,
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a711

Abstract:
Ikan memiliki bentuk dan ukuran tertentu yang berbeda antara ikan yang satu dengan yang lain. Permasalahan dalam mengenali jenis ikan lebih kompleks dibandingkan dengan mengenali wajah manusia. Perbedaan bentuk, warna, dan tekstur pada ikan lebih bervariasi dibandingkan manusia. Pengenalan jenis ikan pada umumnya masih dilakukan secara manual menggunakan pengamatan mata. Sehingga diperlukan adanya sistem yang dapat mengenali ikan secara otomatis. Penelitian sebelumnya juga sudah dapat mengenali jenis ikan namun sensitive terhadap berbagai transformasi atau deformasi dari sebuah objek, dan waktu komputasi yang tidak sedikit, sehingga kurang efektif untuk mengenali objek ikan. Dalam Penelitian ini, kami mengusulkan metode untuk mendeteksi dan mengenali jenis objek ikan menggunakan metode ORB dan KNN. Pengaplikasian dari metode ORB diterapkan untuk ekstraksi fitur dari gambar yang diambil. Kemudian hasil tersebut akan diklasifikasi menggunakan KNN untuk menentukan label kelas yang tepat dari input data ikan. Hasil uji coba menunjukkan bahwa metode yang diusulkan pada penelitian ini mencapai akurasi klasifikasi sebesar 97,5%.
Nursanti Novi Arisa,
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a661

Abstract:
Leukemia is one of the dangerous diseases that can cause death. One of the types of leukemia is acute leukemia that includes ALL (Acute Lymphoblastic Leukemia) and AML (Acute Myeloid Leukemia). The fastest identification against this disease can be done by computing and analysing white blood cell types. However, the manual counting and identification of the white blood cell types are still limited by time. Therefore, automatic counting process is necessary to be conducted in order to get the results more quickly and accurately. Previous studies showed that automatic counting process in the image of Acute Leukemia cells faced some obstacles, the existence of touching cell and the implementation of  geometry feature that cannot produce an accurate counting. It is because the shapes of the cell are various. This study proposed a method for the counting of white blood cells and the separation of touching cells on Acute Leukemia cells image by using Multi Pass Voting method (MPV) based on seed detection (centroid) and K-Means method. Initial segmentation used for separating foreground and background area is canny edge detection. The next stage is seed detection (centroid) using Multi Pass Voting method. The counting of white blood cells is based on the results of the centroid produced. The existence of the touching cells are  separated using K-Means method, the determination of the initial centroid  is based on the results of the Multi Pass Voting method. Based on the evaluation results of 40 images of Acute Leukemia dataset, the proposed method is capable to properly compute based on the centroid. It is also able to separate the touching cell into a single cell. The accuracy of the white blood cell counting result is about 98,6%.
Husnul Hakim, Alexius Reinaldo Budiman
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a732

Abstract:
Penelusuran Minat dan Kemampuan (PMDK) merupakan salah satu jalur penerimaan mahasiswa baru di perguruan tinggi. Berbeda dengan jalur lain yang menggunakan tes tertulis, jalur PMDK merupakan jalur penerimaan mahasiswa baru tanpa melalui tes. Mahasiswa baru akan diseleksi dengan memperhatikan nilai rapor calon mahasiswa selama duduk di bangku SMA. Pada penelitian ini, akan dikembangkan metode seleksi mahasiswa baru melalui jalur PMDK. Calon mahasiswa tidak hanya diseleksi berdasarkan nilai rapor, tetapi juga berdasarkan kualitas sekolah dan histori nilai mahasiswa yang berasal dari sekolah asal pendaftar PMDK. Ketiga parameter ini dapat saling bertentangan. Sebagai contoh, sekolah dengan kualitas yang baik dapat saja memiliki standar yang tinggi sehingga nilai rapor siswanya lebih rendah dari nilai rapor siswa yang berasal dari sekolah lain yang kualitasnya lebih rendah. Untuk itu, perlu digunakan metode pengambilan keputusan yang melibatkan banyak kriteria. Permasalahan pengambilan keputusan seperti ini dikenal dengan multicriteria decision making (MCDM).Salah satu cara untuk pengambilan keputusan MCDM adalah dengan menggunakan Fuzzy TOPSIS. Pada penelitian ini, ketiga parameter yang menentukan diterima atau tidaknya calon mahasiswa akan diproses dengan menggunakan FUZZY TOPSIS. Hasil penelitian menunjukkan bahwa closeness coefficient yang dihasilkan melalui Fuzzy TOPSIS berkolerasi dengan nilai IPK mahasiswa yang diterima melalui jalur PMDK. Kata kunci: Fuzzy TOPSIS, MCDM, PMDK
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a750

Abstract:
Transition region based image segmentation is one of the simple and effective image segmentation methods. This method is capable to segment image contains single or multiple objects. However, this method depends on the background. It may produce a bad segmentation result if the gray level variance is high or the background is textured. So a method to repair the transition region is needed. In this study, a new method to repair the transition region with median filter based on the percentage of the adjacent transitional pixels is proposed. Transition region is extracted from the grayscale image. Transition region refinement is conducted based on thepercentage of the adjacent transitional pixels. Then, several morphological operations and the edge linking process are conducted to the transition region. Afterward, region filling is used to get the foreground area. Finally, image of segmentation result is obtained by showing the pixels of grayscale image that are located in the foreground area.The value of misclassification error (ME), false negative rate (FNR), and false positive rate (FPR) of the segmentation result are calculated to measure the proposed method performance. Performance of the proposed method is compared with the other method. The experimental results show that the proposed method has average value of ME, FPR, and FNR: 0.0297, 0.0209, and 0.0828 respectively. It defines that the proposed method has better performance than the other methods. Furthermore, the proposed method works well on the image with a variety of background, especially on image with textured background.
Dwi Ratnasari, Hayatulloh Firman Hadi, Jian Budiarto
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a738

Abstract:
Futsal lease nowadays is mostly managed manually. The consumers have to come to the site to check the available schedule and to lease it. It is not effective since it spent times. As the technology advances, an android-based application for futsal lease is alternative solution. The design and development of android-based application for futsal lease use waterfall method which started with design, analysis, development, tryout and implementation. The application provides real time information about futsal court, a search feature to available futsal schedules, futsal leases, and the down payment directly in the application. Based on the results of tryout through questionnaire filled by futsal owners (3 subjects) and consumers (17 subjects), 69,7% of the respondents very agree on the use of the application. Thus, the research concludes that android-based application for futsal lease has satisfied the need of futsal owners and consumers.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a754

Abstract:
K-Most Promising (K-MPP) product is a strategy for selecting a product that used in the process of determining the most demanded products by consumers. The basic computations used to perform K-MPP are two types of skyline queries: dynamic skyline and reverse skyline. K-MPP selection is done on the application layer, which is the last layer of the OSI model. One of the application layer functions is providing services according to the user's preferences.In the K-MPP implementation, there exists the situation in which the manufacturer may be less satisfied with the query results generated by the database search process (why-not question), so they want to know why the database gives query results that do not match their expectations. For example, manufacturers want to know why a particular data point (unexpected data) appears in the query result set, and why the expected product does not appear as a query result. The next problem is that traditional database systems will not be able to provide data analysis and solution to answer why-not questions preferred by users.To improve the usability of the database system, this study is aiming to answer why-not K-MPP and providing data refinement solutions by considering user feedback, so users can also find out why the result set does not meet their expectations. Moreover, it may help users to understand the result by performing analysis information and data refinement suggestion.
JUTI: Jurnal Ilmiah Teknologi Informasi, Volume 16; https://doi.org/10.12962/j24068535.v16i2.a762

Abstract:
Sebuah sistem penyaringan otomatis dansistem diagnosa yang akurat sangat berguna untuk proses analisis hasil pemeriksaan pap smear. Langkah yang paling utama dari sistem tersebut adalah proses segmentasi selnukleus dan sitoplasma pada citra hasil pemeriksaan pap smear, karena dapat memengaruhi keakuratan sistem.Normally Biased Generalized Gradient Vector Flow Snake (NBGGVFS) merupakan sebuah algoritma gaya eksternal untuk active contour (snake) yang menggabungkan metode Generalized Gradient Vector Flow Snake (GGVFS) dan Normally Biased Gradient Vector Flow Snake (NBGVFS). Dalam memodelkan snake, terdapat fungsi edge map. Edge map biasanya dihitung dengan menggunakan operator deteksi tepi seperti sobel. Namun, metode ini tidak dapat mendeteksi daerah nukleus dari citra smear serviks dengan benar. Penelitian ini bertujuan untuk segmentasi citra sel tunggal smear serviks dengan memanfaatkan penggunaan Radiating Edge Map untuk menghitung edge map dari citra dengan metode NBGGVFS. Metode yang diusulkan terdiri atas tiga tahapan utama, yaitu tahap praproses, segmentasi awal dan segmentasi kontur. Uji coba dilakukan dengan menggunakan data set Herlev. Pengujian dilakukan dengan membandingkan hasil segmentasi metode yang diusulkan dengan metode pada penelitian sebelumnya dalam melakukan segmentasi citra sel tunggal smear serviks. Hasil pengujian menunjukkan bahwa metode yang diusulkan mampu mendeteksi area nukleus lebih optimal metode penelitian sebelumnya. Nilai rata-rata akurasi dan Zijdenbos Similarity Index (ZSI) untuk segmentasi nukleus adalah 96,96% dan 90,68%. Kemudian, nilai rata-rata akurasi dan ZSI untuk segmentasi sitoplasma adalah 86,78% and 89,35%. Dari hasil evaluasi tersebut, disimpulkan metode yang diusulkan dapat digunakan sebagai proses segmentasi citra smear serviks pada identifikasi kanker serviks secara otomatis.
Page of 7
Articles per Page
by
Show export options
  Select all
Back to Top Top