View Item 
        •   Utrecht University Student Theses Repository Home
        • UU Theses Repository
        • Theses
        • View Item
        •   Utrecht University Student Theses Repository Home
        • UU Theses Repository
        • Theses
        • View Item
        JavaScript is disabled for your browser. Some features of this site may not work without it.

        Browse

        All of UU Student Theses RepositoryBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

        Johnson-Lindenstrauss Transforms

        Thumbnail
        View/Open
        Thesis_Fedor_Taggenbrock_6815057.pdf (1.735Mb)
        Publication date
        2025
        Author
        Taggenbrock, Fedor
        Metadata
        Show full item record
        Summary
        The original Johnson-Lindenstrauss (JL) theorem says that a random data independent matrix can be used to embed high dimensional data into a space of much lower dimension, such that the Euclidean distances between all points are preserved up to an error term with high probability. In this thesis a version of the JL theorem will be proven which states that multiplication by a random matrix consisting of sub- Gaussian random variables can be used to embed the data as described above. The proof relies on some cornerstone results of high dimensional probability theory. This thesis will therefore start with a thorough analysis of properties and concentration inequalities for (squared) sub-Gaussian random variables, such as Hoeffding’s and Bernstein’s inequality. After proving the JL theorem using these results, we will focus on the Fast Johnson-Lindestrauss Transform (FJLT). This result discovered by Ailon and Chazelle in 2006, states that by using a fast preconditioning of the data, a sparse random matrix can be used to obtain the same distance preservation as with the original JL theorem. The benefit of the allowed sparsity is a significant gain in time complexity. Furthermore this result will also be shown to hold for the ℓ1 distances of the data. The proof will again have concentration inequalities at it’s core, but will also rely on convex optimization theory.
        URI
        https://studenttheses.uu.nl/handle/20.500.12932/48749
        Collections
        • Theses
        Utrecht university logo