Skip to content

Commit f837e37

Browse files
committed
link add
1 parent 59b3618 commit f837e37

File tree

3 files changed

+13
-12
lines changed

3 files changed

+13
-12
lines changed

_bibliography/pubs.bib

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,8 @@ @misc{hoover-etal-2022-amlap
44
author = {Hoover, Jacob Louis and Sonderegger, Morgan and Piantadosi, Steven T. and O'Donnell, Timothy J.},
55
year = {2022},
66
howpublished = {Poster at Architectures and Mechanisms for Language Processing (AMLaP 28)},
7-
day = 6,
8-
month = sep,
7+
day = {6},
8+
month = {September},
99
address = {{York, England}},
1010
url = {https://virtual.oxfordabstracts.com/#/event/3067/submission/297},
1111
poster = {https://d3ijlhudpq9yjw.cloudfront.net/cdd9fcfe-abaa-4d50-b0a0-9723ab9e9bc9.pdf}
@@ -15,7 +15,7 @@ @article{hoover-etal-2023-plausibility
1515
title = {The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing},
1616
author = {Hoover, Jacob Louis and Sonderegger, Morgan and Piantadosi, Steven T. and O'Donnell, Timothy J.},
1717
year = {2023},
18-
month = jul,
18+
month = {July},
1919
journal = {Open Mind: Discoveries in Cognitive Science},
2020
volume = {7},
2121
pages = {350--391},
@@ -57,7 +57,7 @@ @inproceedings{hoover-etal-2021-emnlp
5757
author = {Hoover, Jacob Louis and Du, Wenyu and Sordoni, Alessandro and O{'}Donnell, Timothy J.},
5858
booktitle = {Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing},
5959
code = {https://github.com/mcqll/cpmi-dependencies},
60-
month = nov,
60+
month = {November},
6161
pages = {2941--2963},
6262
poster = {2021.10.11.EMNLP.poster.pdf},
6363
publisher = {Association for Computational Linguistics},
@@ -72,9 +72,10 @@ @thesis{hoover.j:2024phd
7272
author = {Hoover, Jacob Louis},
7373
type = {PhD},
7474
year = {2024},
75-
month = aug,
75+
month = {August},
7676
note = {PhD Thesis, McGill University, Linguistics Department},
7777
school = {McGill University},
78-
langid = {english},
79-
url = {https://jahoo.github.io/assets/dissertation.pdf}
78+
langid = {en-CA},
79+
url = {https://escholarship.mcgill.ca/concern/theses/r494vr42w},
80+
pdf = {https://jahoo.github.io/assets/dissertation.pdf}
8081
}

_config.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ scholar:
2727
order: descending
2828
# group_by: year
2929
# group_order: descending
30-
bibtex_skip_fields: [ "abstract", "month_numeric", "poster", "handout", "pdf", "code", "slides" ]
30+
bibtex_skip_fields: ["abstract", "code", "handout", "month_numeric", "myurl", "myurltitle", "openaccess", "pdf", "poster", "preprint", "slides"]
3131
# type_names: { article: Papers }
3232
# linebreaks: true #
3333
source: /_bibliography

_posts/2020-12-22-training-tensor-trains.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
---
22
layout: post
3-
title: Training Tensor Trains
3+
title: A practical comparison of tensor train models
44
comments: true
55
published: true
66
tags:
@@ -13,12 +13,12 @@ I worked on a project with Jonathan Palucci exploring the trainability of a cert
1313
<div style="text-align: center;"><img width="400" src="/assets/2020-12-22-training-tensor-trains-fig2.png"></div>
1414

1515

16-
There is a general correspondence between tensor networks and graphical models, and in particular, when restricted to non-negative valued parameters, Matrix Product States are equivalent to Hidden Markov Models. [Glasser _et al_. 2019](https://arxiv.org/abs/1907.03741) discussed this correspondence, and proved theoretical results about these non-negative models, as well as similar real-- and complex--valued tensor trains. They supplemented their theoretical results with evidence from numerical experiments. In this project, we re-implemented models from their paper, and also implemented time-homogeneous versions of their models.
16+
There is a general correspondence between tensor networks and graphical models, and in particular, when restricted to non-negative valued parameters, [Matrix Product States](https://tensornetwork.org/mps/) are equivalent to Hidden Markov Models (HMMs)). [Glasser _et al_. 2019](https://arxiv.org/abs/1907.03741) discussed this correspondence, and proved theoretical results about these non-negative models, as well as similar real-- and complex--valued tensor trains. They supplemented their theoretical results with evidence from numerical experiments. In this project, we re-implemented models from their paper, and also implemented time-homogeneous versions of their models.
1717
We replicated some of their results for non-homogeneous models, adding a comparison with homogeneous models on the same data. We found evidence that homogeneity decreases ability of the models to fit non-sequential data, but preliminarily observed that on sequential data (for which the assumption of homogeneity is justified), homogeneous models achieved an equally good fit with far fewer parameters. Surprisingly, we also found that the more powerful non time-homogeneous positive MPS performs identically to a time homogeneous HMM.
1818

1919
📊 Poster --> [here (PDF)](/assets/pdfs/2020.12.15.tensor-trains-poster.pdf).
2020

21-
📄 Writeup --> [here (PDF)](/assets/pdfs/2020.12.22.tensor-trains-writeup.pdf).
21+
📄 Writeup titled _A practical comparison of tensor train models: The effect of homogeneity_ --> [here (PDF)](/assets/pdfs/2020.12.22.tensor-trains-writeup.pdf).
2222

23-
💻 Code --> [on GitHub](https://github.com/postylem/tensor_network_project).
23+
💻 Code --> [on GitHub](https://github.com/postylem/comparison-of-tensor-train-models).
2424

0 commit comments

Comments
 (0)