Jekyll2023-11-18T09:25:01-05:00https://paulhe.com/feed.xmlPaul HePaul HeMathematics or Science?2022-12-28T00:00:00-05:002022-12-28T00:00:00-05:00https://paulhe.com/2022/12/28/mathematics-or-science<p>As part of the program chair’s report for POPL 1992, Andrew Appel wrote a paper titled <a href="https://www.cs.princeton.edu/~appel/papers/science.pdf">“Is POPL Mathematics or Science?”</a>
His paper measured how theoretical a conference is using the author ordering of the papers published.
While this paper was meant to “provide some laughs”, I thought it was very interesting and reproduced the results for both the same data and for present-day programming languages conferences.</p>
<p>Most scientific fields usually order by the amount of work each author contributed, sometimes with other conventions, like students being listed first or the grant holder being listed last.
Mathematical fields, however, typically order authors alphabetically, since small ideas are often instrumental in a paper, and it can be difficult to rank the importance of each author’s contributions.
Computer science can go in either direction, depending on the subfield and the paper.
Using the author ordering, we can measure how applied (science) vs. theoretical (math) a conference is.</p>
<h2 id="reproducing-the-1992-results">Reproducing the 1992 Results</h2>
<p>Following Appel’s approach, I recreated the algorithm to compute the most likely proportion of mathematicians to scientists in each conference, with associated error bars.
The code can be found at <a href="https://github.com/Grain/appel">this Git repo</a>.
The data for these conferences was collected from DBLP (copy and pasted, then cleaned up using emacs macros).</p>
<p>The original plot:
<img src="/assets/1992paper.png" alt="Plot of results from the paper" /></p>
<p>My plot, with sloppy manually-added lines:
<img src="/assets/1992mine.png" alt="Plot of results I generated" /></p>
<p>If you look closely at the data, you may notice that many of my results differ from Appel’s original paper.
I had a hard time figuring out the reasons for this.
One possible reason is that sometimes surnames are difficult to discern.
My code uses the <a href="https://github.com/datamade/probablepeople">probablepeople</a> library to guess the surname portion of the name, and sometimes I have to decide when the library can’t figure it out, for example by manually removing whitespace or adding hyphenation in a name.
However, I tried different hyphenation and whitespace combinations for the POPL authors, with no success in reproducing the original paper’s results.
Note that this may still be a factor for other conferences, which I didn’t look at in detail.</p>
<p>My second hypothesis was that the DBLP data differed from the actual author lists that Appel used in 1992.
To test this, I needed a paper copy of the proceedings, since the digital copies on the ACM digital library did not include front matter, and only the individual papers.
I could have looked at each paper individually for the author lists, but I deemed that to be too much work.
Instead, I found a copy of one of the POPL proceedings in the Penn library, and requested for the table of contents to be scanned and sent to me.
A few days later I confirmed that this was the source of the discrepency, at least for that specific data point.
Some of the authors switched to publishing under a different surname, and DBLP displayed their most current name, even for older papers.</p>
<p>To make my work easier, I didn’t do anything to solve these two issues, so my results will differ from Appel’s original results.
For the first, I didn’t do any special handling of surnames, taking the name that probablepeople returned, and only intervening if it could not determine a unique surname.
For the second, I just used the DBLP data as-is.</p>
<h2 id="newer-data">Newer Data</h2>
<p>Next, I ran the algorithm on modern data.
The most difficult part was resolving name issues, when probablepeople could not figure out an author’s surname.
When this happened, I looked up the author and tried to figure out their surname from their personal webpages.
If their surname didn’t really matter, for example when the author list is clearly not alphabetical, I was less careful.
This means that even for the same person, I would edit names inconsistently when it didn’t matter for alphabetical ordering.
Apologies if you’re one of these authors, or if I messed up your name!</p>
<p>One interesting thing I learned while collecting the data is that many of the ACM conferences were consistently held in the US early on.
I was surprised since some of the major conferences currently alternate (well, usually) between North American and non-North-American locations.
I thought this was a tradition since the start, but this has only been happening for about 20 years.
For example, PLDI was in North America for its first 22 iterations.
POPL was in the US until the 14th in Munich, and did not leave the US again until the 24th POPL.</p>
<p>Finally, the results are plotted below.
They suggest that POPL has been getting more practical over time, and that POPL is the most theoretical of the PL conferences, both of which I agree with.</p>
<h3 id="popl">POPL</h3>
<p><img src="/assets/popl.png" alt="Plot of results for all POPL conferences up to 2022" /></p>
<p>The errors bars are far smaller nowadays, owing to the lower number of single-author papers and the higher overall number of papers presented each year.</p>
<h3 id="pl-conferences-20202022">PL Conferences 2020–2022</h3>
<p><img src="/assets/2022.png" alt="Plot of results for some PL conferences from 2020 to 2022" /></p>Paul HeAs part of the program chair’s report for POPL 1992, Andrew Appel wrote a paper titled “Is POPL Mathematics or Science?” His paper measured how theoretical a conference is using the author ordering of the papers published. While this paper was meant to “provide some laughs”, I thought it was very interesting and reproduced the results for both the same data and for present-day programming languages conferences.CV of Failures2021-11-10T00:00:00-05:002023-08-01T00:00:00-04:00https://paulhe.com/2021/11/10/cv-failures<p>Most of what we publicly show of ourselves is positive.
This can give others the impression that they are the odd one out, who has failures when everyone else only seems to succeed.
But in reality, everybody fails—we simply hide the failures.
This CV of failures aims to make my efforts more transparent, and to provide a more balanced perspective on my career.</p>
<p>This idea is not new, but originates from an <a href="https://doi.org/10.1038/nj7322-467a">article</a> by Melanie Stefan.</p>
<h2 id="conference-paper-rejections">Conference Paper Rejections</h2>
<ul>
<li>ESOP 2023</li>
<li>OOPSLA 2022 (both rounds of submission!)</li>
<li>PLDI 2021</li>
<li>ICFP 2019</li>
<li>MFCS 2018</li>
<li>LICS 2018</li>
</ul>
<h2 id="fellowship-and-award-rejections">Fellowship and Award Rejections</h2>
<ul>
<li>Penn CTL Graduate Fellowship 2021-2022</li>
<li>NSERC PGS D 2021</li>
<li>Facebook Fellowship 2021</li>
<li>Penn CTL Graduate Fellowship 2020-2021</li>
<li>NSERC PGS D 2020</li>
<li>NSERC CGS M 2018 at McGill</li>
<li>NSERC CGS M 2018 at uOttawa</li>
<li>UWaterloo K.D. Fryer Medal 2018</li>
</ul>
<h2 id="rejected-school-and-job-applications">Rejected School and Job applications</h2>
<ul>
<li>CMU CS PhD program 2018</li>
<li>Around 100 software engineering internships at UWaterloo, 2014-2016</li>
</ul>
<h2 id="other-rejections">Other rejections</h2>
<ul>
<li>Student volunteering at conferences, all the time</li>
<li>PLDI 2020 SRC, where the top 3 posters win a prize, there were 4 entrants, and I got 4th</li>
</ul>Paul HeMost of what we publicly show of ourselves is positive. This can give others the impression that they are the odd one out, who has failures when everyone else only seems to succeed. But in reality, everybody fails—we simply hide the failures. This CV of failures aims to make my efforts more transparent, and to provide a more balanced perspective on my career.Weak weeks2019-10-23T00:00:00-04:002019-10-23T00:00:00-04:00https://paulhe.com/2019/10/23/weak-weeks<p>In August I tried to abandon the concept of the days of the week.</p>
<p>I had very few scheduled responsibilities: only a single meeting scheduled for each week.
The goal was to reduce mood changes due to the day of the week, like getting no work done on Fridays and disappointment on Sundays.
Though studies on this <sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup> suggest that there is little support for some of these phenomena, I wanted to see for myself.</p>
<p>The first few week were fine; I’d try to go to work around five days a week, and take days off when I felt like I needed one.
I especially tried to work on weekends, since I was more productive (or at least happier with music on my speakers) when the office was empty.
This was a pretty good setup.
I felt like I had a lot of flexibility and got a decent amount done.</p>
<p>But then I got a cat.
I stayed home way too much to watch her, even though she was still adjusting and wasn’t especially playful.
The summer ended with me doing far less than I wanted to.
Nevertheless, I’d like to try this experiment again, especially since it seemed fruitful from the first few weeks.</p>
<p>Anyways, here’s some photos of my cat Strawberry, the real reason for the post.</p>
<p><img src="/assets/strawberry1.jpg" alt="Strawberry" />
<img src="/assets/strawberry2.jpg" alt="Strawberry" />
<img src="/assets/strawberry3.jpg" alt="Strawberry" />
<img src="/assets/strawberry4.jpg" alt="Strawberry" />
<img src="/assets/strawberry5.jpg" alt="Strawberry" />
<img src="/assets/strawberry6.jpg" alt="Strawberry" /></p>
<div class="footnotes" role="doc-endnotes">
<ol>
<li id="fn:1" role="doc-endnote">
<p>Arthur A. Stone, Stefan Schneider & James K. Harter (2012) Day-of-week mood patterns in the United States: On the existence of ‘Blue Monday’, ‘Thank God it’s Friday’ and weekend effects, The Journal of Positive Psychology, 7:4, 306-314, DOI: 10.1080/17439760.2012.691980 <a href="#fnref:1" class="reversefootnote" role="doc-backlink">↩</a></p>
</li>
</ol>
</div>Paul HeIn August I tried to abandon the concept of the days of the week.Coinduction2019-04-17T00:00:00-04:002019-10-23T00:00:00-04:00https://paulhe.com/2019/04/17/coinduction<p>Coinduction is the mathematical dual to an indispensible mathematical tool: induction.
While mathematical induction has been known for thousands of years, coinduction has only been studied for a few decades.
It is still primarily used in computer science, from which it originated in the field of concurrency theory.
Coinduction allows us to define circular or <em>infinite</em> objects (such as <a href="https://en.wikipedia.org/wiki/Stream_(computer_science)">streams</a>, lists that can be infinitely long), and to prove things about them.</p>
<p>It should not be confused with <a href="https://en.wikipedia.org/wiki/Coinduction_(anaesthetics)">this coinduction</a>, which may put you to sleep instead.</p>
<h2 id="inductive-definitions">Inductive definitions</h2>
<p>Inductive (or recursive) definitions are ubiquitous in mathematics, to the point where they are often implicit.
They follow a common pattern to build up a set of objects incrementally.
A base case (or multiple) is first established, and then rules for building up objects based on previous levels are defined.</p>
<p>The set of finite strings \(S\) on an alphabet \(\Sigma\) is the set inductively defined by the following rules, in <a href="https://en.wikipedia.org/wiki/Rule_of_inference">inference rule</a> notation:</p>
\[\frac{}{\epsilon \in S} \qquad \frac{s \in S \quad \sigma \in \Sigma}{\sigma s \in S}\]
<p>So \(\epsilon\) (the empty string) is a string, and for any symbol \(\sigma\) in the alphabet, we can prepend that onto another string to yield a string.
Only the objects generated from the rules are in \(S\).</p>
<p>Inductive definitions can be thought of as an iterative process: we start with the empty set and keep adding objects according to the definition, until in the limit, we reach a <a href="https://en.wikipedia.org/wiki/Fixed_point_(mathematics)">fixed point</a>, when applying the rules no longer adds anything new to the set.
We add \(\epsilon\), then the length 1 strings, then the length 2 strings, and so on, until we have the infinite set of strings over \(\Sigma\) of any length in \(\mathbb{N}\).</p>
<p>An inductive definition is thus the <em>smallest</em> set closed <em>forward</em> under its defining rules.
That is, \(S\) is the smallest set such that \(\epsilon \in S\) and that if \(s \in S\), then \(\sigma s \in S\) for any \(\sigma \in \Sigma\).
We apply the rules from premises to conclusion.</p>
<h2 id="coinductive-definitions">Coinductive definitions</h2>
<p>Since coinduction is the dual to induction, let’s try “flipping” the inductive definition.
A coinductive definition is the <em>largest</em> set closed <em>backward</em> under its defining rules.</p>
<p>What does this mean?
For an inductive definition, we can think of the set as starting from \(\varnothing\) and iteratively adding elements according to the rules.
For a coinductive definition, we can think of the set as starting from the set of all possible objects (even infinite ones), and iterative removing objects that contradict the rules.</p>
<p>If we use the same rules that inductive defined \(S\) above, the coinductively defined set \(S'\) is the largest set such that \(\epsilon \in S'\) and that if \(\sigma s \in S\), then \(s \in S\) (and \(\sigma \in \Sigma\)).
Here, the backward closure goes from the conclusion to the premises, the opposite of the forward closure.
The set of finite strings, \(S\), is included in \(S'\).
But we also have some new strings in \(S'\), the infinitely long strings.
Consider the string \(s = aaaaaa \dots\), where \(a \in \Sigma\).
We cannot construct it using the base case, but it doesn’t lead to a contradiction either, since if \(s = aaaaa \dots \in S\), taking off the first \(a\) results in the same infinite string \(s\), and \(s \in S\) as desired.</p>
<p>The proof tree for \(s\) is infinite, and looks like the following:</p>
\[\large \frac{a \in \Sigma \quad \frac{ a \in \Sigma \quad \frac{ \cdots }{ aaa \dots \in S' } }{ aaa \dots \in S' }}{ aaa \dots \in S' }\]
<p>While objects of inductive definitions require finite derivations, objects of coinductive definitions can have infinite derivations.</p>
<h2 id="proof-principles">Proof principles</h2>
<p>For the following, I will skip over some (many) details.</p>
<p>The function \(F\) can be thought of as the set of rules for a given (co)inductive definition.
\(F(X)\) is the set of conclusions obtained after applying the rules using \(X\) as the set of premises.</p>
<p>Recall that an inductive definition is the least fixed point of a set of rules, and that a coinductive definition is the greatest fixed point.
Now here is a specialization of the Knaster–Tarski fixpoint theorem:</p>
<p><strong>Theorem:</strong> <br />
The least fixed point of \(F = \mu F = \bigcap \{ X \mid F(X) \subseteq X \}\). <br />
The greatest fixed point of \(F = \nu F = \bigcup \{ X \mid X \subseteq F(X) \}\).</p>
<p>\(F(X) \subseteq X\) captures the meaning of the informal “closed forwards” definition from earlier.
Given a set \(T\) where the premises \(X \subseteq T\), we can apply <em>all</em> rules in the “forwards” direction, obtaining the set of conclusions \(F(X)\) which are also in \(T\): \(F(X) \subseteq X \subseteq T\).</p>
<p>Dually, \(X \subseteq F(X)\) captures the meaning of “closed backwards”.
Given a set \(T\) where the conclusions \(F(X) \subseteq T\) from some set of premises \(X\), we can apply <em>some</em> rule for each \(t \in F(X)\) in the “backwards” direction, obtaining the set of premises \(X\) which are also in \(T\): \(X \subseteq F(X) \subseteq T\).</p>
<p>Simple corollaries of the fixpoint theorem gives us proof principles for inductive and coinductive definitions:</p>
<p><strong>Lemma (Induction Principle):</strong> <br />
If \(F(X) \subseteq X\), then \(\mu F \subseteq X\). <br />
<strong>Lemma (Coinduction Principle):</strong> <br />
If \(X \subseteq F(X)\), then \(X \subseteq \nu F\).</p>
<h3 id="proof-by-induction">Proof by induction</h3>
<p>Using the induction principle, we can show that every element of a inductively defined set satisfies some condition, by showing that the condition is preserved for each rule of the definition.</p>
<p>We can derive the more familiar principle of mathematical induction using this.
Let \(F(X) = \{ 0 \} \cup \{ 1 + x \mid x \in X \}\).
This is the set of rules for the natural numbers.
It may be more familiar if I write it as the following:</p>
\[\frac{}{0 \in \mathbb{N}} \qquad \frac{n \in \mathbb{N}}{1 + n \in \mathbb{N}}\]
<p>Then to prove some fact about the natural numbers, we just need to show that it is preserved when applying these rules in the forwards direction.
For example, we will show that \(1 + 2 + \dots + n = \frac{n(n+1)}{2}\) is true for all natural numbers.
Let’s take \(X = \{ n \in \mathbb{N} \mid 1 + 2 + \dots + n = \frac{n(n+1)}{2} \}\).
Then we will prove that \(\mu F = \mathbb{N} \subseteq X\).
This is exactly the conclusion of the Induction Principle, so we need to show that \(F(X) \subseteq X\).</p>
<p>An element of \(F(X)\) can either be \(0\) (the base case), which we can easily verify is in \(X\), or \(1 + n\) (the inductive case) where \(n \in X\) (the inductive hypothesis).
This should look familiar.
Some fiddling will show that the second case is true as well, and we are done! \(\Box\)</p>
<h3 id="proof-by-coinduction">Proof by coinduction</h3>
<p>Dually, using the coinduction principle, we can show that an element is in the coinductively defined set.</p>
<p>Using just \(S'\), our only coinductively defined set so far, would not be very interesting, since it would involve only the membership proofs we saw earlier.
Let’s make another coinductive definition, this time a relation on elements of \(S'\): let \(F(X) = \{ (\epsilon, \epsilon) \} \cup \{ (\sigma_1 s_1, \sigma_2 s_2) \mid \sigma_1 \le \sigma_2 \land (s_1, s_2) \in X \}\), where \(\le\) is some ordering on the alphabet (the usual one on the English alphabet, for instance).
Can you tell what relation this defines?
Let’s write down the inference rules:</p>
\[\frac{}{\epsilon \leqslant \epsilon} \qquad \frac{\sigma_1 \le \sigma_2 \qquad s_1 \leqslant s_2}{\sigma_1 s_1 \leqslant \sigma_2 s_2}\]
<p>The notation should help: \(\nu F\) is the lexicographic ordering relation on our (possibly) infinite strings, displayed here as \(\leqslant\).</p>
<p>Now we can prove that some strings are related by this relation.
For an example, we will show \(aaaa \dots \leqslant baaaa \dots\).
Note that these are infinitely long strings.</p>
<p>Using the coinduction principle, we just need to show that \((aaaa \dots, baaaa \dots)\) is in some set of pairs of strings that is closed backwards under \(F\).
Let’s try the singleton set \(X = \{(aaaa \dots, baaaa \dots)\}\) first.
Then \(F(X) = \{ (\epsilon, \epsilon) \} \cup \{ (\sigma_1 aaaa \dots, \sigma_2 baaaa \dots) \mid \sigma_1 \le \sigma_2 \}\).
But then \(X \not \subseteq F(X)\), since the second string of every pair in \(F(X)\) has a \(b\) as the second symbol.</p>
<p>\(X\) is our “coinductive hypothesis”.
Like how during induction we sometimes have to strengthen the inductive hypothesis, here we have to strengthen the coinductive hypothesis by making it bigger.</p>
<p>Recall the “backwards closed” intuition.
We want to show that by applying <em>some</em> rule “backwards”, we obtain something still in \(X\).
If we start with \((aaaa \dots, baaaa \dots)\), we can only apply the second rule, stripping off the first symbol of each string.
\(a \le b\), so that premise is fine, and we just need to show that \((aaaa \dots, aaaa \dots) \in X\) now.
It looks like we need to grow \(X\) by adding this new pair to it, strengthening the coinductive hypothesis.</p>
<p>Now \(X = \{ (aaaa \dots, baaaa \dots), (aaaa \dots, aaaa \dots) \}\), and \(F(X) = \{ (\epsilon, \epsilon) \} \cup \\ \{ (\sigma_1 aaaa \dots, \sigma_2 baaaa \dots) \mid \sigma_1 \le \sigma_2 \} \cup \\ \{ (\sigma_1 aaaa \dots, \sigma_2 aaaa \dots) \mid \sigma_1 \le \sigma_2 \}\)</p>
<p>Let’s check that \(X \subseteq F(X)\). <br />
\((aaaa \dots, baaaa \dots) = (a\cdot aaaa \dots, b\cdot aaaa \dots)\), and \(a \le b\). <br />
\((aaaa \dots, aaaa \dots) = (a\cdot aaaa \dots, a\cdot aaaa \dots)\), and \(a \le a\).</p>
<p>And since \((aaaa \dots, baaaa \dots) \in X\), we’re done! \(\Box\)</p>
<h2 id="conclusion">Conclusion</h2>
<p>Recently I’ve been working on <a href="https://github.com/DeepSpec/InteractionTrees/">Interaction Trees</a>, a library that provides a coinductive data structure for reasoning about interactive programs in Coq.
Coinduction is less convenient than induction in Coq.
For example, in the coinductive proof above the “coinductive hypothesis” included exactly the conclusion we were trying to prove.
When doing the proof informally, we know we must apply one of the rules backwards and only then can we apply the coinductive hypothesis.</p>
<p>Doing this in a proof assistant like Coq is more complex.
Using “vanilla” Coq, it will allow you to apply the coinductive hypothesis immediately, and then complain that you got it wrong when you try to finish the proof.
The <a href="https://github.com/snu-sf/paco">paco</a> library solves this problem, but more complex reasoning quickly gets complex, which is why I started learning more about the theory behind coinduction.</p>
<p>I find it really intriguing how (relatively) new coinduction is and how useful it has become.
There’s been a lot of work recently on areas related to coinduction, and I’m excited to do more work in this area.</p>
<h2 id="resources">Resources</h2>
<p>I first encountered coinduction in <a href="https://www.cis.upenn.edu/~bcpierce/tapl/">Types and Programming Languages</a> by Benjamin C. Pierce, where they are introduced to talk about the metatheory of recursive types. While I wouldn’t recommend reading this if you’re <em>just</em> interested in coinduction, it serves as an excellent introduction to programming languages and type systems.</p>
<p><a href="http://www.cs.unibo.it/~sangio/IntroBook.html">Introduction to Bisimulation and Coinduction</a> by Davide Sangiorgi is a very accessible textbook that goes into detail about all of this and more.
It cleared up a lot of questions I had about coinduction, and helped me understand it more rigorously.</p>Paul HeCoinduction is the mathematical dual to an indispensible mathematical tool: induction. While mathematical induction has been known for thousands of years, coinduction has only been studied for a few decades. It is still primarily used in computer science, from which it originated in the field of concurrency theory. Coinduction allows us to define circular or infinite objects (such as streams, lists that can be infinitely long), and to prove things about them.Welcome2019-03-06T00:00:00-05:002020-08-18T00:00:00-04:00https://paulhe.com/2019/03/06/first-post<p>Welcome to my blog! This is the first (official) post, hopefully of many.</p>
<h2 id="goals">Goals</h2>
<p>I decided to start this blog because I rarely write anything longer than a sentence at a time, which seems like a useful thing to practice for a PhD student.
I’ve also always felt pretty weak at communicating about research or technical stuff (not to mention just in general).
Hopefully this will help me with these things, as well as improve my understanding of the technical material I’ll be writing about.</p>
<p>I plan to write about various technical things I encounter during my research work.
These will probably be things related to functional programming and programming language theory.</p>
<h2 id="website">Website</h2>
<p>Let me tell you a bit about the inner workings of the website, which I spent (and will continue to spend) a lot of time on—instead of writing posts.
The site is static, hosted on <a href="https://pages.github.com/">Github Pages</a>, and is generated by <a href="https://jekyllrb.com/">Jekyll</a>.
I don’t want to handle any complexity related to hosting, so a static website seems fine to me.</p>
<p>However if you look at the bottom of the page, you’ll see something less standard: an ugly hacked together comment system.
I originally tried using <a href="https://disqus.com/">Disqus</a>, which was really quite nice and easy to use.
You can see an example of it on my blog <a href="/2019/02/28/first-post.html">here</a>.</p>
<p>I preferred something more lightweight though, and also something I controlled entirely.
Here’s an <a href="/2019/02/28/second-post.html">example</a> of what my current solution looks like.
You can even embed html (what’s sanitization?)!</p>
<p>I’m using <a href="https://staticman.net/">Staticman</a> to display user-generated content.
When a comment is submitted, it goes through the Staticman web service, which creates a pull request on my website’s Github repo to add the comment as a text file, to be included in the (updated) static site.</p>
<p>I think this is super cool.
No databases or anything to deal with!
Staticman was pretty nice to use, though the documentation is a little out of date.
Originally it was run as a single public instance, and due to the number of users it had it was hitting the rate limit for the Github API (see <a href="https://github.com/eduardoboucas/staticman/issues/243">this Github issue</a> for details).
The developer then updated Staticman to be a Github app, so each user would get their own instance and thus their own API quota.
However, this was fairly recent (Dec 2018) and the documentation wasn’t updated to reflect this, so it took me a few hours to get it working.</p>
<p><strong>Edit</strong> Oct 3: Staticman died at some point! The service is open source, so maybe I will host my own instance on Heroku. <br />
<strong>Edit</strong> Aug 18 2020: Finally hosted my own instance on Heroku. <br />
<strong>Edit</strong> Dec 21 2022: Heroku shut down their free tier, but I’ve managed to migrate to fly.io’s free tier with basically a single click using their migration tool.</p>
<h2 id="to-come">To Come</h2>
<p>I’ll be working on updating the site a bit more to add a navigation menu (<strong>Edit</strong> Mar 7: Done!), less ugly comments (<strong>Edit</strong> Mar 8: Arguably done!), and so on.
For my research I’ll be spending most of my time in the next few weeks on coinduction, a very cool proof technique, and I hope to write something introductory about it soon.</p>Paul HeWelcome to my blog! This is the first (official) post, hopefully of many.Blogging Like a Hacker2019-02-28T00:00:00-05:002019-02-28T00:00:00-05:00https://paulhe.com/2019/02/28/first-post<h1 id="test">Test</h1>
<div id="disqus_thread"></div>
<script>
/**
* RECOMMENDED CONFIGURATION VARIABLES: EDIT AND UNCOMMENT THE SECTION BELOW TO INSERT DYNAMIC VALUES FROM YOUR PLATFORM OR CMS.
* LEARN WHY DEFINING THESE VARIABLES IS IMPORTANT: https://disqus.com/admin/universalcode/#configuration-variables
*/
/*
var disqus_config = function () {
this.page.url = https://paulhe.com/2019/02/28/first-post.html
this.page.identifier = /2019/02/28/first-post
};
*/
(function() { // DON'T EDIT BELOW THIS LINE
var d = document, s = d.createElement('script');
s.src = 'https://paulhe.disqus.com/embed.js';
s.setAttribute('data-timestamp', +new Date());
(d.head || d.body).appendChild(s);
})();
</script>
<noscript>Please enable JavaScript to view the <a href="https://disqus.com/?ref_noscript" rel="nofollow">comments powered by Disqus.</a></noscript>
<script id="dsq-count-scr" src="//paulhe.disqus.com/count.js" async=""></script>Paul HeTestComments2019-02-28T00:00:00-05:002019-02-28T00:00:00-05:00https://paulhe.com/2019/02/28/second-postPaul He