Streaming data may consist of multiple drifting concepts each having its own underlying data distribution. We present an ensemble learning based approach to handle the data streams having multiple underlying modes. We...
详细信息
Streaming data may consist of multiple drifting concepts each having its own underlying data distribution. We present an ensemble learning based approach to handle the data streams having multiple underlying modes. We build a global set of classifiers from sequential data chunks; ensembles are then selected from this global set of classifiers, and new classifiers created if needed, to represent the current concept in the stream. The system is capable of performing any-time classification and to detect concept drift in the stream. In streaming data historic concepts are likely to reappear so we don't delete any of the historic classifiers. Instead, we judiciously select only pertinent classifiers from the global set while forming the ensemble set for a classification task.
Content matching based algorithms form the core of many network security devices. It is one of the critical components due to the fact that it allows making decisions based on the actual content flowing through the ne...
详细信息
Content matching based algorithms form the core of many network security devices. It is one of the critical components due to the fact that it allows making decisions based on the actual content flowing through the network. The most important parameters that go into the design of a content matching algorithm are its performance and accuracy of detection. Although this topic had received significant attention in literature over past decade, much of the work was focused on improving the performance. The accuracy of detection was limited within a packet instance. Protocols like TCP do not guarantee that message boundaries are preserved. This can result in a segmented pattern across packets. This paper demonstrates a novel flow-aware content matching algorithm that solves this limitation without compromising the performance.
data dependences (dataflow constraints) present a major hurdle to the amount of instruction-level parallelism that can be exploited from a program. Recent work has focused on the use of data value prediction to overc...
详细信息
data dependences (dataflow constraints) present a major hurdle to the amount of instruction-level parallelism that can be exploited from a program. Recent work has focused on the use of data value prediction to overcome the limits imposed by data dependences. That is, when an instruction is fetched, its result can be predicted so that subsequent instructions that depend on the result can execute earlier using the predicted value. When the correct result becomes available, it is compared against the value predicted earlier, so as to validate the prediction. Whereas significant work has been done towards developing schemes for accurately predicting data values, not much work has been done towards understanding and quantifying the performance impact of data value prediction. This paper presents a quantitative study of the impact of data value prediction on available parallelism. Our studies, done with the MIPS instruction set and a collection of SPEC95 integer benchmarks, show that data value prediction provides significant increases in available parallelism when infinite size instruction window and perfect branch prediction are used. Our studies with finite size windows shows that the impact of data value prediction is not very significant for small window sizes such as 64. When the instruction window size is increased, the benefits of data value prediction become more apparent.
data-flow testing relies on static analysis for computing the definition-use pairs that serve as the test case requirements for a program. When testing large programs, the individual procedures are first tested in iso...
详细信息
data-flow testing relies on static analysis for computing the definition-use pairs that serve as the test case requirements for a program. When testing large programs, the individual procedures are first tested in isolation during unit testing. Integration testing is performed to specifically test the procedure interfaces. The procedures in a program are integrated and tested in several steps. Since each integration step requires data-flow analysis to determine the new test requirements, the accumulated cost of repeatedly analyzing a program can contribute considerably to the overhead of testing. data-flow analysis is typically computed using an exhaustive approach or by using incremental data-flow updates. This paper presents a new and more efficient approach to data-flow integration testing that is based on demand-driven analysis. We developed and implemented a demand-driven analyzer and experimentally compared its performance during integration testing with the performance of (i) a traditional exhaustive analyzer, and (ii) an incremental analyzer. Our experiments show that demand-driven analysis is faster than exhaustive analysis by up to a factor of 25. The demand-driven analyzer also outperforms the incremental analyzer in 80% of the test programs by up to a factor of 5.
An approach for reconstruction of sparse high-resolution data from lower-resolution dense spatiotemporal data is introduced. The basic idea is to compute the dense feature velocities from lower-resolution data and pro...
详细信息
ISBN:
(纸本)076951695X
An approach for reconstruction of sparse high-resolution data from lower-resolution dense spatiotemporal data is introduced. The basic idea is to compute the dense feature velocities from lower-resolution data and project them to the corresponding high-resolution data for computing the missing data. In this context, the basic flow equation is solved for intensity, as opposed to feature velocities at high resolution. Although the proposed technique is generic, we have applied our approach to sea surface temperature (SST) data at 18 km (low-resolution dense data)for computing the feature velocities and at 4 km (high-resolution sparse data) for interpolating the missing data. At low resolution, computation of flow field is regularized and uses the incompressibility constraints for tracking fluid motion. At high resolution, computation of intensity is regularized for continuity across multiple frames.
An important issue in the development of high performance computer architectures is how to automatically partition algorithms for parallel execution. The cyclostatic realization method (CRM) is identified as a basis f...
详细信息
An important issue in the development of high performance computer architectures is how to automatically partition algorithms for parallel execution. The cyclostatic realization method (CRM) is identified as a basis for an effective mapping methodology to achieve automatic partitioning. The paper develops an advanced mapping methodology for the block dataflow architecture (BDFA) based on CRM. Weaknesses in the existing methodology are identified and an effective, generalized methodology for correcting these weaknesses is proposed. Several key factors are investigated: a system-wide average iteration period for multiple input streams, the effect of communication overhead, and the influence of network topology. The enhanced mapping methodology is applied to the BDFA for a 2D 2nd order IIR filter.
Consideration of the problem of Satellite Backbone data Network of SCS "Yamal" (SBDN) organisation and definition of position that it takes place in RJSC "Gazprom" Corporative Network (CN).
Consideration of the problem of Satellite Backbone data Network of SCS "Yamal" (SBDN) organisation and definition of position that it takes place in RJSC "Gazprom" Corporative Network (CN).
The most straight forward approach to requirements elicitation is to provide a direct transformation from the user's task descriptions to analysis models. A detailed description of tasks is formalized in ontology....
详细信息
The most straight forward approach to requirements elicitation is to provide a direct transformation from the user's task descriptions to analysis models. A detailed description of tasks is formalized in ontology. Using the task ontology it is possible to create templates to facilitate transforming tasks to both use cases and dataflow diagrams. In this paper, a new user-oriented approach for task analysis is presented to decrease process in analysis systems.
This paper presents the design of a Videophone Coder-Decoder Motion Estimator using two High-Level Synthesis tools. Indeed, the combination of a Control flow Dominated part (complex communication protocol) with a data...
详细信息
This paper presents the design of a Videophone Coder-Decoder Motion Estimator using two High-Level Synthesis tools. Indeed, the combination of a Control flow Dominated part (complex communication protocol) with a dataflow Dominated part (high throughput computations) makes this circuit difficult to be synthesized by a single HLS tool. The combination of two HLS tools for the high-level design of this operator required the definition of a sophisticated design flow allowing mixed-level and multi-language simulations. When compared to design starting from RTL specifications, HLS induces only a negligible area overhead of 5%, and provides gain in description length (divided by 5), design time and flexibility.
暂无评论