摘要:
In a method for visualizing query results in stream processing systems, a visualization service receives a query from a client to visualize data in a stream processing application. The query is sent from the visualization service to a query-able operator of the stream processing application. At the query-able operator, an operation is performed using history data in the query-able operator to produce a first result that satisfies the query and the first result is sent to the visualization service. At the query-able operator, another operation is performed using new data received by the query-able operator to produce a second result that satisfies the query and the second result is sent to the visualization service. The first and second results are output from the visualization service to the client.
摘要:
A method and computer program product for assembling and deploying multi-platform flow based applications. An information processing flow that produces a result is assembled, the information processing flow includes components connected by data links, a component includes software code that describes at least one of an input constraint or an output constraint of the component, and at least two of the components are deployable on different computing platforms. The information processing flow is partitioned into sub-flows, such that for each sub-flow every component in the sub-flow is deployable on the same computing platform. The sub-flows are deployed on their respective computing platforms.
摘要:
In a method for visualizing query results in stream processing systems, a visualization service receives a query from a client to visualize data in a stream processing application. The query is sent from the visualization service to a query-able operator of the stream processing application. At the query-able operator, an operation is performed using history data in the query-able operator to produce a first result that satisfies the query and the first result is sent to the visualization service. At the query-able operator, another operation is performed using new data received by the query-able operator to produce a second result that satisfies the query and the second result is sent to the visualization service. The first and second results are output from the visualization service to the client.
摘要:
An apparatus, including a memory device for storing a program, and a processor in communication with the memory device, the processor operative with the program to facilitate design of a stream processing flow that satisfies an objective, wherein the stream processing flow includes at least three processing groups, wherein a first processing group includes a data source and an operator, a second processing group includes a data source and an operator and a third processing group includes a join operator at its input and another operator, wherein data inside each group is organized by channels and each channel is a sequence of data, wherein an operator producing a data channel does not generate new data for the channel until old data of the channel is received by all other operators in the same group, and wherein data that flows from the first and second groups to the third group is done asynchronously and is stored in a queue if not ready for processing by an operator of the third group, and deploy the stream processing flow in a concurrent computing system to produce an output.
摘要:
An apparatus to facilitate design of a stream processing flow that satisfies an objective, wherein the flow includes at least three processing groups, wherein a first processing group includes a data source and an operator, a second processing group includes a data source and an operator and a third processing group includes a join operator at its input and another operator, wherein data inside each group is organized by channels and each channel is a sequence of data, wherein an operator producing a data channel does not generate new data for the channel until old data of the channel is received by all other operators in the same group, and wherein data that flows from the first and second groups to the third group is done asynchronously and is stored in a queue if not ready for processing by an operator of the third group.
摘要:
A method and computer program product for assembling and deploying multi-platform flow based applications. An information processing flow that produces a result is assembled, the information processing flow includes components connected by data links, a component includes software code that describes at least one of an input constraint or an output constraint of the component, and at least two of the components are deployable on different computing platforms. The information processing flow is partitioned into sub-flows, such that for each sub-flow every component in the sub-flow is deployable on the same computing platform. The sub-flows are deployed on their respective computing platforms.
摘要:
A microfluidic filtration unit for trapping particles of a predetermined nominal size present in a fluid is provided. The unit comprises a fluid chamber connected to an inlet for introducing the fluid to be filtered and an outlet for discharging filtered fluid, a filtration barrier arranged within the fluid chamber, said filtration barrier comprising a plurality of pillars arranged substantially perpendicular to the path of fluid flow when fluid is introduced into the fluid chamber, said pillars being aligned to form at least one row extending across said path of fluid flow, wherein each of said at least one row of pillars in the filtration barrier comprises at least one fine filtration section comprising a group of pillars that are spaced apart to prevent particles to be filtered from the fluid from moving between adjacent pillars, and at least one coarse filtration section comprising a group of pillars that are spaced apart to permit the movement of particles between adjacent pillars.
摘要:
Methods and apparatus operating in a stream processing network perform load shedding and dynamic resource allocation so as to meet a pre-determined utility criterion. Load shedding is envisioned as an admission control problem encompassing source nodes admitting workflows into the stream processing network. A primal-dual approach is used to decompose the admission control and resource allocation problems. The admission control operates as a push-and-pull process with sources pushing workflows into the stream processing network and sinks pulling processed workflows from the network. A virtual queue is maintained at each node to account for both queue backlogs and credits from sinks. Nodes of the stream processing network maintain shadow prices for each of the workflows and share congestion information with neighbor nodes. At each node, resources are devoted to the workflow with the maximum product of downstream pressure and processing rate, where the downstream pressure is defined as the backlog difference between neighbor nodes. The primal-dual controller iteratively adjusts the admission rates and resource allocation using local congestion feedback. The iterative controlling procedure further uses an interior-point method to improve the speed of convergence towards optimal admission and allocation decisions.
摘要:
A silicon condenser microphone is described. The silicon condenser microphone of the present invention comprises a perforated backplate comprising a portion of a single crystal silicon substrate, a support structure formed on the single crystal silicon substrate, and a floating silicon diaphragm supported at its edge by the support structure and lying parallel to the perforated backplate and separated from the perforated backplate by an air gap.
摘要:
Systems, methods, and products for pull data transfer in a request-response model are provided herein. One aspect provides for generating output data utilizing at least one data generation station; and communicating via the at least one data generation station output data related to at least one data request received from at least one data requesting station responsive to at least one criterion, the at least one criterion comprising one of expiration of a time period or generation of a threshold amount of output data. Other embodiments and aspects are also described herein.