Princeton University users: to view a senior thesis while away from campus, connect to the campus network via the Global Protect virtual private network (VPN). Unaffiliated researchers: please note that requests for copies are handled manually by staff and require time to process.
 

Publication:

SpectraLDS: Distilling Spectral Filters into Constant-Time Recurrent Models

datacite.rightsrestricted
dc.contributor.advisorHazan, Elad
dc.contributor.authorFortgang, Shlomo T.
dc.date.accessioned2025-08-06T15:43:45Z
dc.date.available2025-08-06T15:43:45Z
dc.date.issued2025-04-10
dc.description.abstractWe introduce the first provable method for learning a symmetric linear dynamical system of arbitrarily high effective memory. This allows us to distill the convolutional layers in a leading hybrid state space model, FlashSTU, into O(1) linear dynamical systems, merging Transformer and RNN architectures in a manner suitable for scaling and with application to language modeling and other sequential processing tasks.
dc.identifier.urihttps://theses-dissertations.princeton.edu/handle/88435/dsp01z890rx705
dc.language.isoen_US
dc.titleSpectraLDS: Distilling Spectral Filters into Constant-Time Recurrent Models
dc.typePrinceton University Senior Theses
dspace.entity.typePublication
dspace.workflow.startDateTime2025-04-10T22:36:05.888Z
pu.certificateOptimization and Quantitative Decision Science
pu.contributor.authorid920245534
pu.date.classyear2025
pu.departmentComputer Science

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
written_final_report.pdf
Size:
1.07 MB
Format:
Adobe Portable Document Format
Download

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
100 B
Format:
Item-specific license agreed to upon submission
Description:
Download