Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01j098zd77v
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorKevrekidis, Ioannis G-
dc.contributor.authorHoliday, Alexander-
dc.contributor.otherChemical and Biological Engineering Department-
dc.date.accessioned2017-12-12T19:18:08Z-
dc.date.available2017-12-12T19:18:08Z-
dc.date.issued2017-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/dsp01j098zd77v-
dc.description.abstractRecent decades have seen a tremendous rise in the affordability and performance of various computational technologies, enabling researchers to propose and probe ever more complicated numerical models. These simulations often generate incredible quantities of data that must be sifted through to glean useful conclusions. This thesis highlights our efforts to automate this process in two specific areas: (a) uncovering simplified descriptions of dynamic network models and (b) detecting important parameter combinations in general nonlinear systems. Both advances involve modification of the manifold learning algorithm, Diffusion Maps (DMAPS), to address the particular problem. In the first case, the challenge is quantifying the similarity of two networks in a reasonable amount of computational time. We propose a number of possible solutions, and examine their performance when combined with DMAPS. We find that by combining suitable measures of similarity with DMAPS we are able to uncover low-dimensional structure in a set of networks, thus enabling us to describe the system in terms of one or two values instead of thousands. In the second, we must extend Diffusion Maps to operate on the graph of a function. In particular we consider a model that maps parameter values to some output. By properly formulating the DMAPS kernel, we enable DMAPS to discover the directions in parameter space along which model predictions vary most significantly. Both result in algorithms that we hope are practically useful to researchers in a variety of fields who are looking for simplified descriptions of their complex systems.-
dc.language.isoen-
dc.publisherPrinceton, NJ : Princeton University-
dc.relation.isformatofThe Mudd Manuscript Library retains one bound copy of each dissertation. Search for these copies in the library's main catalog: <a href=http://catalog.princeton.edu> catalog.princeton.edu </a>-
dc.subjectdiffusion maps-
dc.subjectdimensionality reduction-
dc.subjectmanifold learning-
dc.subjectnetworks-
dc.subjectparameter reduction-
dc.subject.classificationApplied mathematics-
dc.titleManifold learning for coarse-graining networks and for parameter reduction-
dc.typeAcademic dissertations (Ph.D.)-
pu.projectgrantnumber690-2143-
Appears in Collections:Chemical and Biological Engineering

Files in This Item:
File Description SizeFormat 
Holiday_princeton_0181D_12351.pdf24.25 MBAdobe PDFView/Download


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.