Serwis Infona wykorzystuje pliki cookies (ciasteczka). Są to wartości tekstowe, zapamiętywane przez przeglądarkę na urządzeniu użytkownika. Nasz serwis ma dostęp do tych wartości oraz wykorzystuje je do zapamiętania danych dotyczących użytkownika, takich jak np. ustawienia (typu widok ekranu, wybór języka interfejsu), zapamiętanie zalogowania. Korzystanie z serwisu Infona oznacza zgodę na zapis informacji i ich wykorzystanie dla celów korzytania z serwisu. Więcej informacji można znaleźć w Polityce prywatności oraz Regulaminie serwisu. Zamknięcie tego okienka potwierdza zapoznanie się z informacją o plikach cookies, akceptację polityki prywatności i regulaminu oraz sposobu wykorzystywania plików cookies w serwisie. Możesz zmienić ustawienia obsługi cookies w swojej przeglądarce.
With the acceleration of the Internet speed and the vigorous emergence of new applications, the amount of Internet traffic has increased. In order to provide stable Internet service, efficient network management based on accurate traffic identification is critical. Although various methods for traffic identification have been proposed, not a single method identifies all types of Internet traffic....
IP traceback to the origin of denial of service (DoS) attack is extremely hard and most existing methods such as marking and logging based approaches are not practical and need changes in infrastructure. The flow-based IP traceback is a new approach which can use the advantages of metaheuristic algorithms. In this paper, a bee colony algorithm is applied to find the DoS attack source. The traffic...
Along with the information explosion in the Internet era, the traditional classification methods, such as KNN (k-nearest neighbor), Naive Bayes (NB), encounter bottlenecks due to the endless stream of new words. In this paper, through comparing with the Rocchio and Bayesian algorithms, it has been found that centroid-based algorithms are insufficient for text classification. Therefore, a novel feature...
Real-time internet traffic classification has potential applications in next-generation internet security and bandwidth management. Current machine learning-based algorithms for traffic classification, however, present scalability issues that would degrade system performance if executed to make control decisions on real-time streams. This tension gives rise to competing performance costs for traffic...
The use of online review sites has grown significantly, allowing for communities to share information on products or services.These online review sites are marketed as being independent and trustworthy, but have been criticised for not ensuring the integrity of the reviews.One major concern is that of review fraud; where a person (such as a marketer) is paid to write favourable reviews for one product...
Content-Centric Network (CCN) provides a cleanslate design for the Internet, where content becomes the primitive of communications. In CCN, routers are equipped with content stores, which act as caches for frequently requested content. This design enables the Internet to provide content distribution services without any application-layer support. On the other hand, as caches are integrated into routers,...
This paper introduces a topic specific web crawling system, which gathers basic educational resources from the web, and indexes them for the purpose of basic educational users. Compared to other similar theme based crawling system, the crawler integrates fuzzy rule based algorithm and VSM text analysis technology together to predicting each URL's relevancy to basic education while parsing current...
Internet is becoming an increasingly important platform for ordinary life and work. It is expected that keyword extraction can help people quickly find hot spots on the web, since keywords in a document provide important information about the content of the document. In this paper, we propose to use text clustering method based on semi-supervised learning to get focuses of social topics in a large...
The web is a comprehensive and dynamic repository of information regarding most of the areas of human knowledge. However, retrieving information from the web is not an easy task. Web's characteristics place many difficulties to users willing to explore it as an information source. Information retrieved from the web is usually very extensive, composed of voluminous collections of hypertext documents...
Tabu Search is a meta-heuristic approach successfully used to address optimization problems in several contexts. This paper reports the results of an empirical study carried out to investigate the effectiveness of Tabu Search in estimating Web application development effort. The dataset employed in this investigation is part of the Tukutuku database. This database has been used in several studies...
In a variety of domains, the amount of information grows rapidly; new sources and types of information are proliferating. In recent years, the world-wide web information has been growing at a dramatic pace and might have become the most important information source for most people. In daily life, when people encounter some problems, they tend to retrieve the information from the Web search engines...
As computer systems become increasingly complex, system anomalies have become major concerns in system management. In this paper, we present a comprehensive measurement study to quantify the predictability of different system anomalies. Online anomaly prediction allows the system to foresee impending anomalies so as to take proper actions to mitigate anomaly impact. Our anomaly prediction approach...
Accurately locating the geographical position of Internet hosts has many useful applications. Existing approaches for host geolocation use Internet latency measurements, IP-to-location mapping and also geographical and demographical hints. In this paper, we investigate the applicability of the Maximum Likelihood Estimation (MLE) technique for Internet host geolocation. Our approach is based on a probability...
The automatic allocation of enterprise workload to resources can be enhanced by being able to make `what-if' response time predictions, whilst different allocations are being considered. It is important to quantitatively compare the effectiveness of different prediction techniques for use in cloud infrastructures. To help make the comparison of relevance to a wide range of possible cloud environments...
Similarity calculation has many applications, such as information retrieval, and collaborative filtering, among many others. It has been shown that link-based similarity measure, such as SimRank, is very effective in characterizing the object similarities in networks, such as the Web, by exploiting the object-to-object relationship. Unfortunately, it is prohibitively expensive to compute the link-based...
This paper presents the first work known publicly on spoken term detection from bilingual spontaneous speech using code-switched lattice-based structures for word and subword units. The corpus used is the lectures with Chinese as the host language and English as the guest language recorded for a real course offered in National Taiwan University. The techniques reported here have been successfully...
Predicting latency between nodes on the Internet can have a significant impact on the performance of many services that use latency distances among nodes as a decision making input. Coordinate-based approaches are among the family of latency prediction techniques where latency between each pair of nodes is modeled as the virtual distance among those nodes over a virtual system. This article proposes...
Enormous adverse impacts of Internet worms have attracted strong interest from the academia to model their propagation. This paper presents a dynamic parameter branching process to characterize the spread of scanning worm on the Internet. Firstly, the paper demonstrates that the spread of scanning worm is a typical branching process with dynamic parameters rather than a static parameter. Secondly,...
This paper explores the effect of profile based method for classification of Wikipedia XML documents. Our approach builds two profiles, exploiting the whole content, Initial Descriptions and links in the Wikipedia documents. For building profiles we use the negative category information which has shown to perform well for classifying unstructured texts. The performance of Cosine and Fractional Similarity...
Current technology allows steganography applications to conceal any digital file inside of another digital file. Due to the large number of steganography tools available over the Internet, a particular threat exists when criminals use steganography to conceal their activities within digital images in cyber space. In this paper, a set of statistical features are generated using linear mixed effects...
Podaj zakres dat dla filtrowania wyświetlonych wyników. Możesz podać datę początkową, końcową lub obie daty. Daty możesz wpisać ręcznie lub wybrać za pomocą kalendarza.