The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
A main advantage of app stores is that they aggregate important information created by both developers and users. In the app store product pages, developers usually describe and maintain the features of their apps. In the app reviews, users comment these features. Recent studies focused on mining app features either as described by developers or as reviewed by users. However, extracting and matching...
User profiling is a typical big data service created and utilized by an increasing number of Internet venders, which maintains a customized model of interests or essential attributes of their existing users by looking for insights into their behaviors. The Internet industry's best practices indicate that user profiles can help venders much more sufficiently understand their customers. As a result,...
GitHub is the largest collaborative source code hosting site built on top of the Git version control system. The availability of a comprehensive API has made GitHub a target for many software engineering and online collaboration research efforts. In our work, we have discovered that a) obtaining data from GitHub is not trivial, b) the data may not be suitable for all types of research, and c) improper...
Nowadays, in the period of the digitalization and knowledge economy development, majority of activities result in the increase of data that should be captured. In each area of business, there is an increasing urge to extract the knowledge from data in a timely manner in order to be able to make shifts that ensure a competitive advantage. Thus, the knowledge of methods and techniques of big data processing...
In the emerging field of big data, a large volume of data has to be managed, operating on data of huge volume becomes easier when it's sorted and structured. The data can be structured using a simple algorithm i.e. index algorithm which stores and categories data on basis of their application. This in turn will be very beneficial on business level as well as on software level.
Collaboration in business processes and projects requires a division of responsibilities among the participants. Version control systems allow us to collect profiles of the participants that hint at participants' roles in the collaborative work. The goal of this paper is to automatically classify participants into the roles they fulfill in the collaboration. Two approaches are proposed and compared...
This paper presents a software toolkit that can be used to analyze event data streams in real-time. It has a specific focus on stochastic analysis of business processes, based on event data that is produced during the execution of those processes. The toolkit provides a software environment that facilitates easy connection to event data streams and quick development and testing of analysis and visualization...
In recent years, with the gradual development of mobile Internet technology, the number of mobile applications increases dramatically. Users facing numerous mobile applications are often caught off guard. It is necessary to automatically classify the applications according to the applications' information, so as to recommend appropriate applications to users. However, the text information directly...
In this paper, we present a powerful end-to-end data mining system that collects application related data and provides insightful relevant fields analysis in addition to search and filtering. We present details on field extraction, indexing, relevant field processing and dynamic baseline derivation. We also propose to demonstrate the effectiveness of various scoring algorithms. Two real-world use...
The first step of a technological solution for the recovery of non-technical energy losses (NTL) in the distribution phase is to use data mining techniques. MVM Ingeniería de Software has developed a technology with high potential, oriented in this direction, for Companies Distributors and Traders of energy (DT), through its advanced analytical capacity in descriptive and predictive levels framed...
Does gamification work? This paper examines how Stack Overflow users behave when earning badges. A regression analysis of user activity logs shows users change their contribution amounts when earning some badges but not others. This paper adds new support to the growing literature that gamification works, but its efficacy is context-dependent. Alternative methods for motivating user contributions...
In this paper, we take the guide data and the program data from the users of digital cable television programs as the experimental data to carry on our experiment. And we use the SAS software platform which is very efficient in data analyzing to realize the classification of our user to different parts. So we can achieve our destination of personal recommendation and precision advertizing.
Human needs of information increase periodically. These improvements lead to knowledge discovery which can be used to prediction. One of the organization that needs prediction to support their business processes is college or university. Universities need to predict their student achievement as support of their business objective. In fact, the software owned by universities have not been able to help...
Process mining is an emerging research area that aims to improve the analysis of Business Process Models (BPMs) by extracting knowledge from event logs. What actually happened in the organization is set forth for consideration, not what people think about the organization. Therefore, it can be used in various industrial and scientific applications. This paper aims to provide information about the...
Due to high competition in today's business and the need for satisfactory communication with customers, companies understand the inevitable necessity to focus not only on preventing customer churn but also on predicting their needs and providing the best services for them. The purpose of this article is to predict future services needed by wireless users, with data mining techniques. For this purpose,...
For IT Infrastructure Support (ITIS), it is crucial to identify opportunities for reducing service costs and improving service quality. We focus on streamlining service levels i.e., finding right resolution level for each ticket, to reduce time, efforts and cost for ticket handling, without affecting workloads and user satisfaction. We formalize this problem and present two statistics-based search...
Even though data warehousing (DW) requires huge investments, the data warehouse market is experiencing incredible growth. However, a large number of DW initiatives end up as failures. In this paper, we argue that the maturity of a data warehousing process (DWP) could significantly mitigate such large-scale failures and ensure the delivery of consistent, high quality, “single-version of truth” data...
XBRL is the latest technology used in processing accounting information. The core part of XBRL is the creation and application of taxonomy. XBRL taxonomy engineering is a special case of software engineering. From the software engineering perspective, the result of the design phase in XBRL taxonomy is the metadata model and the result of the building phase is XBRL element definition converted from...
We present, in this paper, a proposal for the improvement of the CRISP-DM data mining methodology. The first phase of CRISP-DM is focused on the business process and its objectives. This process is made in an informal way, leaving to the analyst the responsibility for funding the entire process. By this reason, we propose the incorporation of two software engineering diagrams, pre-conceptual schemas...
Concept lattice is a new mathematical tool for data analysis and knowledge processing. Attribute reduction is very important in the theory of concept lattice because it can make the discovery of implicit knowledge in data easier and the representation is simple. Knowledge discovery has received more and more attention from the business community for the last few years. One of the most important and...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.