Skip to content

Commit 4f839dc

Browse files
committed
update publications and info
1 parent 20aebca commit 4f839dc

11 files changed

Lines changed: 30 additions & 18 deletions

File tree

_bibliography/papers.bib

Lines changed: 14 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,15 +20,27 @@ @article{keyu2025explanations
2020
journal={ACL Findings},
2121
abstract={Language models today are widely used in education, yet their ability to tailor responses for learners with varied informational needs and knowledge backgrounds remains under-explored. To this end, we introduce ELI-WHY, a benchmark of 13.4K "Why" questions to assess the pedagogical capabilities of LLMs. We then conduct two extensive human studies to assess the utility of LLM-generated explanatory answers (explanations) on our benchmark, tailored to three distinct educational grades: elementary, high-school, and graduate school. In our first study, human raters assume the role of an "educator" to assess model explanations' fit to different educational grades. We find that GPT-4-generated explanations match their intended educational background only 50% of the time, compared to 79% for human-curated explanations. In our second study, human raters assume the role of a learner to assess if an explanation fits their own informational needs. Results show that users deemed GPT-4-generated explanations relatively 20% less suited to their informational needs, particularly for advanced learners. Additionally, automated evaluation metrics reveal that GPT-4 explanations for different informational needs remain indistinguishable in their grade-level, limiting their pedagogical effectiveness. These findings suggest that LLMs' ability to follow inference-time instructions alone is insufficient for producing high-utility explanations tailored to users' informational needs.},
2222
selected={true},
23+
doi={10.48550/arXiv.2506.14200},
2324
pdf={ELI_Why_Evaluating_the_Pe.pdf}
2425
}
2526

2627
@article{keyu2025vlm,
27-
title={Beyond the Text: How Explanation Qualities Influence User Trust in Visual Language Models},
28+
title={Believing without Seeing: Quality Scores for Contextualizing Vision-Language Model Explanations},
2829
author={Keyu He and Tejas Srinivasan and Brihi Joshi and Xiang Ren and Jesse Thomason and Swabha Swayamdipta},
2930
year={2025},
3031
journal={Submitted to NeurIPS, Under Review},
3132
abstract={When people query Vision-Language Models (VLMs) but cannot see the accompanying visual context (e.g. for blind and low-vision users), augmenting VLM predictions with natural language explanations can signal which model predictions are reliable. However, prior work has found that explanations can easily convince users that inaccurate VLM predictions are correct. To remedy undesirable overreliance on VLM predictions, we propose evaluating two complementary qualities of VLM-generated explanations via two quality scoring functions. We propose Visual Fidelity, which captures how faithful an explanation is to the visual context, and Contrastiveness, which captures how well the explanation identifies visual details that distinguish the model’s prediction from plausible alternatives. On the A-OKVQA and VizWiz tasks, these quality scoring functions are better calibrated with model correctness than existing explanation qualities. We conduct a user study in which participants have to decide whether a VLM prediction is accurate without viewing its visual context. We observe that showing our quality scores alongside VLM explanations improves participants’ accuracy at predicting VLM correctness by 11.1%, including a 15.4% reduction in the rate of falsely believing incorrect predictions. These findings highlight the utility of explanation quality scores in fostering appropriate reliance on VLM predictions.},
32-
selected={true}
33+
selected={true},
34+
pdf={VLM_Rationales.pdf}
3335
}
3436

37+
@article{enhancing-debugging,
38+
title = {Enhancing Debugging Skills of LLMs with Prompt Engineering},
39+
author = {Keyu He* and Max Li* and Joseph Liu*},
40+
abstract = {This paper presents a comprehensive study on improving the debugging capabilities of Large Language Models (LLMs) like GPT-3.5, focusing on the application of prompt engineering techniques. We explore the efficacy of few-shot learning, chain-of-thought prompting, and a baseline zero-shot model in enhancing LLMs' ability to debug code. Utilizing static and dynamic evaluation metrics, the study rigorously assesses the debugging proficiency of these models. By introducing different types of bugs, including procedural and language model-generated errors, and applying varied prompting strategies, we provide a deeper understanding of LLMs' debugging capabilities. The results provide insights into the limitation of debugging capabilities of GPT-3.5 Turbo, even with the assistance of various prompting techniques. Source code of our evaluation method and bug generation techniques are in GitHub repository: https://github.com/FrankHe2002/CSCI499FinalProject.},
41+
journal = {Tech Report},
42+
selected = {false},
43+
year = {2023},
44+
month = dec,
45+
pdf = {Enhancing-Debugging.pdf}
46+
}

_config.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ footer_text: >
1414
Hosted by <a href="https://pages.github.com/" target="_blank">GitHub Pages</a>.
1515
keywords: jekyll, jekyll-theme, academic-website, portfolio-website # add your own keywords or leave empty
1616
lang: en # the language of your site (for example: en, fr, cn, ru, etc.)
17-
icon: prof_pic.jpg # the emoji used as the favicon (alternatively, provide image name in /assets/img/)
17+
icon: keyu.jpg # the emoji used as the favicon (alternatively, provide image name in /assets/img/)
1818

1919
url: https://Keyu-He.github.io # the base hostname & protocol for your site
2020
baseurl: # the subpath of your site, e.g. /blog/. Leave blank for root
@@ -102,7 +102,7 @@ spotify_id: # your spotify id
102102
stackoverflow_id: # your stackoverflow id
103103
telegram_username: # your Telegram user name
104104
unsplash_id: # your unsplash id
105-
wechat_qr: wechat-qr.jpg # filename of your wechat qr-code saved as an image (e.g., wechat-qr.png if saved to assets/img/wechat-qr.png)
105+
wechat_qr: wechat_qr.jpg # filename of your wechat qr-code saved as an image (e.g., wechat_qr.png if saved to assets/img/wechat_qr.png)
106106
whatsapp_number: # your WhatsApp number (full phone number in international format. Omit any zeroes, brackets, or dashes when adding the phone number in international format.)
107107
wikidata_id: # your wikidata id
108108
wikipedia_id: # your wikipedia id (Case sensitive)

_pages/about.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -6,28 +6,28 @@ permalink: /
66

77
profile:
88
align: right
9-
image: prof_pic.jpg
9+
image: keyu.jpg
1010
image_circular: true # crops the image to make it circular
11-
# more_info: >
12-
# <p>555 your office number</p>
13-
# <p>123 your address street</p>
14-
# <p>Your City, State 12345</p>
11+
more_info: >
12+
<p>6404 Gates & Hillman Centers</p>
13+
<p>4902 Forbes Ave</p>
14+
<p>Pittsburgh, PA 15213</p>
1515
1616
news: true # includes a list of news items
1717
selected_papers: true # includes a list of papers marked as "selected={true}"
1818
social: true # includes social icons at the bottom of the page
1919
---
2020

21-
Hi! I am Keyu, a senior undergraduate student double majoring in **Computer Science** and **Applied & Computational Mathematics** at the University of Southern California ([USC](https://www.usc.edu/)), with a minor in **Artificial Intelligence Applications**.
21+
Hi! I’m Keyu He, a Master of Intelligent Information Systems ([MIIS](https://miis.cs.cmu.edu/)) student at Carnegie Mellon University ([CMU](https://www.cmu.edu/)). Previously, I earned my B.S. from the University of Southern California ([USC](https://www.usc.edu/)), where I double‑majored in **Computer Science** and **Applied & Computational Mathematics** and minored in **Artificial Intelligence Applications**.
2222

23-
Passionate about Natural Language Processing (NLP), I’m fascinated by its potential to bridge communication gaps and foster understanding. I am currently working on NLP research as a member of the [INK Lab](https://inklab.usc.edu/) and [DILL Lab](https://dill-lab.github.io/), under the guidance of Professors [Xiang Ren](https://www.seanre.com/) and [Swabha Swayamdipta](https://swabhs.com/), and the mentorship of [Brihi Joshi](https://brihijoshi.github.io/). My work focuses on **explainability and building user trust** in AI systems.
23+
Passionate about Natural Language Processing ([NLP](https://en.wikipedia.org/wiki/Natural_language_processing)), I’m fascinated by its potential to bridge communication gaps and foster understanding. I worked on NLP research as a member of the [INK Lab](https://inklab.usc.edu/) and [DILL Lab](https://dill-lab.github.io/), under the guidance of Professors [Xiang Ren](https://www.seanre.com/) and [Swabha Swayamdipta](https://swabhs.com/), and the mentorship of [Brihi Joshi](https://brihijoshi.github.io/). My work focuses on **explainability and building user trust** in AI systems.
2424

2525
### Academic Achievements & Honors
2626
- **USC Center for Undergraduate Research in Viterbi Engineering (CURVE) fellow**.
2727
- Fellowship awarded multiple times (Spring 2024 -- Spring 2025), with a total funding amount of $6,750.
2828
- **USC Academic Achievement Award**: Fall 2022 - Fall 2024.
2929
- This award covered 11 units of tuition costs in total, amounting to approximately $24,000.
30-
- **USC Dornsife Dean’s List & USC Viterbi Dean's List**: 2021-2024.
30+
- **USC Dornsife Dean’s List & USC Viterbi Dean's List**: 2021-2025.
3131
- **4th Place, USC Integral Bee Competition**: 2022.
3232

3333
### Current Research

_pages/profiles.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,15 +10,15 @@ nav_order: 7
1010
# # if you want to include more than one profile, just replicate the following block
1111
# # and create one content file for each profile inside _pages/
1212
# - align: right
13-
# image: prof_pic.jpg
13+
# image: keyu.jpg
1414
# content: about_einstein.md
1515
# image_circular: false # crops the image to make it circular
1616
# more_info: >
1717
# <p>555 your office number</p>
1818
# <p>123 your address street</p>
1919
# <p>Your City, State 12345</p>
2020
# - align: left
21-
# image: prof_pic.jpg
21+
# image: keyu.jpg
2222
# content: about_einstein.md
2323
# image_circular: false # crops the image to make it circular
2424
# more_info: >

_posts/2024-01-27-advanced-images.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,6 @@ This is a simple image slider. It uses the [Swiper](https://swiperjs.com/) libra
3030
This is a simple image comparison slider. It uses the [img-comparison-slider](https://img-comparison-slider.sneas.io/) library. Check the [examples page](https://img-comparison-slider.sneas.io/examples.html) for more information of what you can achieve with it.
3131

3232
<img-comparison-slider>
33-
{% include figure.liquid path="assets/img/prof_pic.jpg" class="img-fluid rounded z-depth-1" slot="first" %}
34-
{% include figure.liquid path="assets/img/prof_pic_color.png" class="img-fluid rounded z-depth-1" slot="second" %}
33+
{% include figure.liquid path="assets/img/keyu.jpg" class="img-fluid rounded z-depth-1" slot="first" %}
34+
{% include figure.liquid path="assets/img/keyu_color.png" class="img-fluid rounded z-depth-1" slot="second" %}
3535
</img-comparison-slider>

assets/img/wechat-qr.jpg

-224 KB
Binary file not shown.

assets/img/wechat_qr.jpg

210 KB
Loading

assets/pdf/VLM_Rationales.pdf

3.31 MB
Binary file not shown.

lighthouse_results/desktop/alshedivat_github_io_al_folio_.html

Lines changed: 1 addition & 1 deletion
Large diffs are not rendered by default.

0 commit comments

Comments
 (0)