Page Not Found
Sitemap
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Pages
Archive Layout with Content
Posts by Category
Posts by Collection
CV
Markdown
Page not in menu
Page Archive
Projects
Publications
Sitemap
Posts by Tags
Talk map
Talks and presentations
Teaching
Terms and Privacy Policy
Blog posts
Jupyter notebook markdown generator
Posts
Future Blog Post
This post will show up by default. To disable scheduling of future posts, edit config.yml and set future: false.
Blog Post number 4
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Blog Post number 3
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Blog Post number 2
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Blog Post number 1
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
portfolio
Recognizing Underlying Concepts in Student Explanations
In this project, I’m exploring computational tools to understand student utterances. In learning, students come up with unique explanations for phenomena that they encounter. These explanations are constructed by piecing together a range of different and relevant concepts. Understanding how the underlying concepts change can give us a window to observe, in fine detail, whena and how learning happens. In this project, I explore the use of language model embeddings to computationally dissect student utterances in interviews where they attempt to explain “Why is it hotter in the summer and colder in the winter?”. I found that, while noisy, embeddings derived from language models (BERT, GPT, etc.) contain information relevant to underlying concepts (e.g. “The Earth spins.”) and can be teased out. More work needs to be done in understanding the structure of the embedding space and applying this methods to larger scale datasets.
publications
Dissecting the Ullman Variations with a SCALPEL: Why do LLMs fail at Trivial Alterations to the False Belief Task?
Why might LLMs succeed on traditional false belief tasks but fail when variations are made? Are they exclusively relying on spurious correlations learned from the training set and don’t really have any ToM capabilities? Or are there other explanations?
talks
Talk 1 on Relevant Topic in Your Field
This is a description of your talk, which is a markdown files that can be all markdown-ified like any other post. Yay markdown!
Tutorial 1 on Relevant Topic in Your Field
Talk 2 on Relevant Topic in Your Field
Conference Proceeding talk 3 on Relevant Topic in Your Field
This is a description of your conference proceedings talk, note the different field in type. You can put anything in this field.
teaching
Teaching experience 1
This is a description of a teaching experience. You can use markdown like any other post.
Teaching experience 2
This is a description of a teaching experience. You can use markdown like any other post.
