Publications
Search results
Jump to search filtersEmploying Sparse Quadrature for Stochastic Optimization in Power Grids
Abstract not provided.
Surrogate-based model for optimization under uncertainty
Abstract not provided.
ESCAPE: Efficiently Counting All 5-Vertex Subgraphs
Abstract not provided.
Counter-Adversarial Community Detection: Initial Investigations
Abstract not provided.
Enabling adaptive scientific workflows via trigger detection
Proceedings of ISAV 2015: 1st International Workshop on In Situ Infrastructures for Enabling Extreme-Scale Analysis and Visualization, Held in conjunction with SC 2015: The International Conference for High Performance Computing, Networking, Storage and Analysis
Next generation architectures necessitate a shift away from traditional workflows in which the simulation state is saved at prescribed frequencies for post-processing analysis. While the need to shift to in situ workflows has been acknowledged for some time, much of the current research is focused on static workflows, where the analysis that would have been done as a post-process is performed concurrently with the simulation at user-prescribed frequencies. Recently, research efforts are striving to enable adaptive workflows, in which the frequency, composition, and execution of computational and data manipulation steps dynamically depend on the state of the simulation. Adapting the workflow to the state of simulation in such a data-driven fashion puts extremely strict efficiency requirements on the analysis capabilities that are used to identify the transitions in the workflow. In this paper we build upon earlier work on trigger detection using sublinear techniques to drive adaptive workflows. Here we propose a methodology to detect the time when sudden heat release occurs in simulations of turbulent combustion. Our proposed method provides an alternative metric that can be used along with our former metric to increase the robustness of trigger detection. We show the effectiveness of our metric empirically for predicting heat release for two use cases.
MaxOutProbe: An Algorithm for Increasing the Size of Partially Observed Networks
Abstract not provided.
Counting Triangles in Real-World Graph Streams: Dealing with Repeated Edges and Time Windows
Abstract not provided.
Modeling Large-Scale Networks
Abstract not provided.
Final Report: Sublinear Algorithms for In-situ and In-transit Data Analysis at Exascale
Post-Moore's law scaling is creating a disruptive shift in simulation workflows, as saving the entirety of raw data to persistent storage becomes expensive. We are moving away from a post-process centric data analysis paradigm towards a concurrent analysis framework, in which raw simulation data is processed as it is computed. Algorithms must adapt to machines with extreme concurrency, low communication bandwidth, and high memory latency, while operating within the time constraints prescribed by the simulation. Furthermore, in- put parameters are often data dependent and cannot always be prescribed. The study of sublinear algorithms is a recent development in theoretical computer science and discrete mathematics that has significant potential to provide solutions for these challenges. The approaches of sublinear algorithms address the fundamental mathematical problem of understanding global features of a data set using limited resources. These theoretical ideas align with practical challenges of in-situ and in-transit computation where vast amounts of data must be processed under severe communication and memory constraints. This report details key advancements made in applying sublinear algorithms in-situ to identify features of interest and to enable adaptive workflows over the course of a three year LDRD. Prior to this LDRD, there was no precedent in applying sublinear techniques to large-scale, physics based simulations. This project has definitively demonstrated their efficacy at mitigating high performance computing challenges and highlighted the rich potential for follow-on re- search opportunities in this space.
An Efficient Approach for Stochastic Optimization of Electricity Grid Operations
Abstract not provided.
Counter-Adversarial Data Analytics: Machine Learning (and Graph Analysis)
Abstract not provided.
Diamond Sampling for Approximate Maximum All-pairs Dot-product (MAD) Search
Abstract not provided.
A Dashboard for Comparative Visualization of Many-variate Data
Abstract not provided.
Path Sampling: A Fast and Provable Method for Estimating 4- vertex subgraph counts
Abstract not provided.
Graph Characterization and Sampling Algorithms
Abstract not provided.
An Efficient Approach for Stochastic Optimization of Electricity Grid Operations
Abstract not provided.
Computing the Largest Entries in a Matrix Product via Sampling
Abstract not provided.
Resiliency of the Power Grid
Abstract not provided.
Toward using surrogates to accelerate solution of stochastic electricity grid operations problems
2014 North American Power Symposium, NAPS 2014
Stochastic unit commitment models typically handle uncertainties in forecast demand by considering a finite number of realizations from a stochastic process model for loads. Accurate evaluations of expectations or higher moments for the quantities of interest require a prohibitively large number of model evaluations. In this paper we propose an alternative approach based on using surrogate models valid over the range of the forecast uncertainty. We consider surrogate models based on Polynomial Chaos expansions, constructed using sparse quadrature methods. Considering expected generation cost, we demonstrate that the approach can lead to several orders of magnitude reduction in computational cost relative to using Monte Carlo sampling on the original model, for a given target error threshold.
Sampling and Streaming Algorithms for Counting Small Patterns in BIG Graphs
Abstract not provided.
Directed closure measures for networks with reciprocity
Journal of Complex Networks
Abstract not provided.
Finding Hierarchical and Overlapping Dense Subgraphs using Nucleus Decompositions
Finding dense substructures in a graph is a fundamental graph mining operation, with applications in bioinformatics, social networks, and visualization to name a few. Yet most standard formulations of this problem (like clique, quasiclique, k-densest subgraph) are NP-hard. Furthermore, the goal is rarely to nd the \true optimum", but to identify many (if not all) dense substructures, understand their distribution in the graph, and ideally determine a hierarchical structure among them. Current dense subgraph nding algorithms usually optimize some objective, and only nd a few such subgraphs without providing any hierarchy. It is also not clear how to account for overlaps in dense substructures. We de ne the nucleus decomposition of a graph, which represents the graph as a forest of nuclei. Each nucleus is a subgraph where smaller cliques are present in many larger cliques. The forest of nuclei is a hierarchy by containment, where the edge density increases as we proceed towards leaf nuclei. Sibling nuclei can have limited intersections, which allows for discovery of overlapping dense subgraphs. With the right parameters, the nuclear decomposition generalizes the classic notions of k-cores and k-trusses. We give provable e cient algorithms for nuclear decompositions, and empirically evaluate their behavior in a variety of real graphs. The tree of nuclei consistently gives a global, hierarchical snapshot of dense substructures, and outputs dense subgraphs of higher quality than other state-of-theart solutions. Our algorithm can process graphs with tens of millions of edges in less than an hour.
Counting Triangles in Real-World Graph Streams: Dealing with Repeated Edges and Time Windows
Abstract not provided.
Strategies for Probing Incomplete Graphs
Abstract not provided.