The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
User profiling is a typical big data service created and utilized by an increasing number of Internet venders, which maintains a customized model of interests or essential attributes of their existing users by looking for insights into their behaviors. The Internet industry's best practices indicate that user profiles can help venders much more sufficiently understand their customers. As a result,...
Despite the globalization of software development, relevant documentation of a project, such as requirements and design documents, often still is missing, incomplete or outdated. However, parts of that documentation can be found outside the project, where it is fragmented across hundreds of textual web documents like blog posts, email messages and forum posts, as well as multimedia documents such...
The paper describes approach, practical solution and intermediate results for self-driven exploration of personal interests structure and nearest social environment for an individual in order to correct their own behavioral patterns and interactions with communication partners on social networks on the Internet.
While big data analytics continue to grow in popularity among companies and organizations, their large-scale analytic implementations are often completed by software developers with little or no formal training in machine learning or data analysis. These developers are skilled at writing code but they do not have the understanding of the data analytics process to be efficient or necessarily accurate...
Because of the rapid growth of open source software, how to choose software from many alternatives becomes a great challenge. Traditional ranking approaches mainly focus on the characteristics of the software themselves. In this paper we investigate the market demands for software engineers, and propose a novel approach for ranking software by analyzing the market requirements for special software...
Public software repositories offer a great opportunity for researchers. GitHub is a repository with more than 10 million projects. GitHub has an implementation of a defect tracking system. This paper describes the process developed to extract defects from GitHub repository, one of the most widely used public repositories. In this work, besides of the process, it is presented the appeared difficulties,...
GitHub is a social coding platform that enables developers to efficiently work on projects, connect with other developers, collaborate and generally "be seen: by the community. This visibility also extends to prospective employers and HR personnel who may use GitHub to learn more about a developer's skills and interests. We propose a pipeline that automatizes this process and automatically suggests...
Abstract- Many engineers are exposed to binary data. These can be files or data exchanged over network links. When involved in the verification and validation of systems that deal with specific protocols or binary data storage, it is often tedious to analyse the hexadecimal dumps in order to find specific parameters of interest. Despite detailed protocol specifications, it takes a lot of manual effort...
The gateway to the internet starts for almost 66% of the world's population at a search engine. Most of us use it as the starting point for that e-commerce purchase, flight reservation, hotel reservation, ordering pizza, obtaining relevant news, research articles and the list is endless. We are able to do all this now, with a few clicks and some amount of recollection from memory. All this, hoping...
A project's documentation is the primary source of information for developers using that project. With hundreds of thousands of programming-related questions posted on programming Q&A websites, such as Stack Overflow, we question whether the developer-written documentation provides enough guidance for programmers. In this study, we wanted to know if there are any topics which are inadequately...
In this paper we will discuss several approaches that can be used to develop an oracle in chatting software. This oracle is meant to give early warning to the chatting software user about their chatting partner based on their communication. The approaches include expert rule, data mining, ontology as well as simple text processing. Even though the issue of privacy and free-information in cyber world...
This paper presents an application of Natural Language Tool (NLT) to support the VPRG extraction of text based vulnerability description. The NLT is used to analyze the text-based vulnerability descriptions to retrieve vulnerability properties and evaluate their relationships. Then, a graph based VPRG model that describes the vulnerability can be established. Finally, with fine-tuning from domain...
Nowadays applications are increasingly developed based on remote Web services and service composition has become a powerful novel development paradigm. Due to the fact that such applications in the context of Internet are deployed and executed in an open and dynamic environment, adaptability is one of the crucial requirements for developing such applications. In our previous work, we presented a variability-based...
To elicit software requirements, we have to have knowledge about a problem domain, e.g., healthcare, shopping or banking where the software is applied. A description of domain knowledge such as a domain ontology helps requirements analysts to elicit requirements completely and correctly to some extent even if they do not have such knowledge sufficiently. Several requirements elicitation methods and...
In order to improve the veracity of the web search, this paper studies the PageRank algorithm, proposes a new method PBTP Algorithm (PageRank Based on Transition Probability), that is an improvement for the classical PageRank method. As forwarding links in a web page are different, the transition probability of a link to be clicked is different too. For the classical PageRank value, should assign...
Online social networks are popular habitats for many Web users. Research on activity patterns of individual features of online social networking systems is ongoing. Our approach on the study of these patterns is more encompassing than previous efforts. We have created a modern experimental educational online social network for the purpose of the study of network structures and communication phenomena...
To describe the social network in online forum (BBS) of university, social network analysis and data mining method was used to investigate the network relationship of community under the help of UCINET software and ID3 algorithm. Characteristic of community network, formation of opinion leader's position and the relationship between initiative and other attributes of community members were also explored...
This paper presents a new conceptual framework and practical solution for Customer Relationship Management (CRM) and E-Loyalty programs for cutting edge M-Commerce. CRM has come to the world of technology to help companies maximise technology usage. CRM means disciplined business strategy to create and sustain long-term, profitable customer relationships. To this aim, it must concentrate on customer...
Web data mining is the integration of data mining technology and web, how to find useful information in the Web which is the world's largest data of collection is becoming a hot spot of data mining research, this article choosing all the 763 documents' citation as data sample of “web data mining(WDM)”, which was published in Web of Science (SCI-EXPANDED, SSCI, A&HCI) from 1986 to 2009, confirming...
Botnet was composed of the virus-infected computers severely threaten the security of internet. Hackers, firstly, implanted virus in targeted computers, which were then commanded and controlled by them via the internet to operate distributed denial of services (DDoS), steal confidential information, distribute junk mails and other malicious acts. By imitating P2P software, P2P botnet used multiple...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.