Rose Profile Banner
Rose Profile
Rose

@rose_e_wang

1,564
Followers
242
Following
66
Media
268
Statuses

NLP & Education @stanfordnlp 🌲 Prev: 2020 MIT 🦫, Google Brain 🧠, Google Brain Robotics 🤖

Joined January 2012
Don't wanna be here? Send us removal request.
Pinned Tweet
@rose_e_wang
Rose
7 days
What do great teachers do to be good at teaching? What can this teach us about LMs? Most of us experience the “front stage” of teaching—as students. Few see the *back stage*: the planning, pedagogical decisions… 🌉 Bridge, at NAACL’24, surfaces these hidden decisions🧵
Tweet media one
3
16
65
@rose_e_wang
Rose
2 years
I’m excited to share “Language modeling via stochastic processes”, an #ICLR2022 oral! Our work addresses the challenge of generating long _coherent_ sequences with language models by leveraging goal-conditioned latent paths. Paper: 🧵[1/T]
Tweet media one
8
110
663
@rose_e_wang
Rose
4 years
Excited to share gym-cooking, a *novel multi-agent Gym environment*: Based on recent work (, #CogSci2020 computational modeling prize winner) with amazing collaborators Sarah Wu, James Evans, Josh Tenenbaum, David Parkes, @maxkw !
6
60
305
@rose_e_wang
Rose
2 years
How can we algorithmically figure out what our model doesn’t know, and then construct datasets to improve it? We tackle this question in “Know thy student: Interactive learning with Gaussian processes” at #ICLR2022 @cells2societies workshop. Paper: [1/N]
Tweet media one
1
23
192
@rose_e_wang
Rose
2 months
Students listening to lectures can go to Google to answer their questions. But…how can the teacher find what part of their lecture *caused* the student’s question in the first place?? 🤔 Introducing *Backtracing*: Retrieving the Cause of the Query! 🧵
Tweet media one
8
37
188
@rose_e_wang
Rose
4 months
There’s a *lot* of insight about how students learn in education data like classroom conversations, but I can tell you it's *painful* to process and analyze. Introducing *Edu-ConvoKit*, a pipeline that handles the pre-processing, annotation, and analysis for you! 🧵
Tweet media one
2
19
95
@rose_e_wang
Rose
3 years
How do we train language models (LMs) to be good pragmatic conversational partners? We investigate this in our #EMNLP2021 Findings paper: Calibrate your listeners! Robust communication-based training for pragmatic speakers. 📜: 📺:
2
9
60
@rose_e_wang
Rose
2 months
#EACL2024 is over!! This was a super fun conference & it was great meeting new folks 😃 Malta is so beautifully colorful --- here are some of my favorite pics!! ✨🎨
Tweet media one
Tweet media two
Tweet media three
Tweet media four
2
2
57
@rose_e_wang
Rose
7 months
Ever wonder how experienced math teachers & tutors compare to ChatGPT or GPT4 in teaching students? 🖥️🧑‍🎓👩‍🏫 Check out our new paper “Step-by-Step Remediation of Students’ Mathematical Mistakes”! 📜 🖥️ from @stanfordnlp @StanfordEd
2
22
51
@rose_e_wang
Rose
10 months
Can ChatGPT help teachers by providing effective feedback, like generating helpful pedagogical suggestions? 👩‍🏫 We answer this question in our work presented @ BEA (co-hosted @ ACL) on Thursday July 13 Harbour A! Website: w/ @ddemszky !
Tweet media one
1
12
47
@rose_e_wang
Rose
2 years
Come check out “Language modeling via stochastic processes” at #ICLR2022 on this Monday! 😄 Looking forward to meeting old and new friends 🥳 Livestream: Apr 25 5-5:15pm PT Poster: Apr 25 6:30-8:30pm PT w/ @esindurmusnlp , Noah Goodman & @tatsu_hashimoto
@rose_e_wang
Rose
2 years
I’m excited to share “Language modeling via stochastic processes”, an #ICLR2022 oral! Our work addresses the challenge of generating long _coherent_ sequences with language models by leveraging goal-conditioned latent paths. Paper: 🧵[1/T]
Tweet media one
8
110
663
0
7
44
@rose_e_wang
Rose
3 years
It's finally time! Join us tomorrow (Friday) for a series of exciting discussions on the roadmap to never-ending RL with the amazing @danijarh @aharutyu @joelbot3000 @pyoudeyer @NataliaDiazRodr @AleksandraFaust @hyogweon and more! 🥳🥳🥳 Schedule:
@neverendingrl
Never Ending RL
3 years
1/N We are excited to introduce our @iclr_conf workshop: A Roadmap to Never-Ending RL. We invite you to submit papers (up to 6 pages, excluding references and appendix) in the @iclr_conf format. Submission Deadline: February 26, 2021 #NERL2021 #ICLR2021
Tweet media one
1
29
112
1
9
37
@rose_e_wang
Rose
3 years
@Shreyagupta08 I’M HIS PHD STUDENT AND I’VE NEVER SEEN HIM IN PERSON I’M SCREAMINGGGGG
2
0
36
@rose_e_wang
Rose
2 months
In Malta 🇲🇹 for EACL this week!! Let me know if you’re around and wanna talk about nlp applications esp education, or just hang with me @chengmyra1 @krisgligoric … or run with me along the coast!! 😎🏃🏻‍♀️
Tweet media one
0
4
31
@rose_e_wang
Rose
2 months
Presenting Backtracing in Malta 🇲🇹 tomorrow at EACL 14:00 in Session A, Radisson Ballroom Level 2 — come say hi!!! 😃
@rose_e_wang
Rose
2 months
Students listening to lectures can go to Google to answer their questions. But…how can the teacher find what part of their lecture *caused* the student’s question in the first place?? 🤔 Introducing *Backtracing*: Retrieving the Cause of the Query! 🧵
Tweet media one
8
37
188
0
4
29
@rose_e_wang
Rose
22 days
How will ed tech change w LLMs? What is and isn't possible? If these Qs have been on your mind, submit your work to a workshop I'm organizing: Leveraging LLMs for Next Gen Ed Tech @ EDM 2024 by May 10th! ➡️ #EDM #EdTech
2
5
28
@rose_e_wang
Rose
3 years
Curious about how we can build cooperative, human-like AI systems? 🤖 📜: Poster 10-11am PST Saturday #NeurIPS2020 Cooperative AI Workshop! 🗣: Spotlight talk 11:45pm PT Saturday 📚: Paper w/ S. Wu, J. Evans, J. Tenenbaum, D. Parkes, @maxhkw !
1
2
26
@rose_e_wang
Rose
3 years
Woo, my internship project is out on the @GoogleAI blog!!! How can we get real-world robots to anticipate each other’s behavior and collaborate? Find out more below!👇 Big thanks to: J. Chase Kew, Dennis Lee, Tsang-Wei Lee, Tingnan Zhang, @brian_ichter , Jie Tan, @AleksandraFaust
@GoogleAI
Google AI
3 years
Introducing a model-based #RL approach for robot navigation, called hierarchical predictive planning (HPP), that enables agents to align their goals on the fly in order to solve the decentralized rendezvous task. Learn more at
7
197
818
0
1
26
@rose_e_wang
Rose
3 years
Excited to present #CoRL2020 paper on model-based RL for multirobot coordination! Work done at #Google w J. Chase Kew, Dennis Lee, Tsang-Wei Lee, Tingnan Zhang, @brian_ichter , Jie Tan, @AleksandraFaust .😄 Website: Live session: Tomorrow (11/18) 12:30pmPT
0
2
25
@rose_e_wang
Rose
8 days
📢 Calling the #EdTech community! Intrigued by the [potential/positive/negative] impact of LLMs on education? Submit your work to this workshop at #EDM2024 🪇👩‍🏫 ➡️ Deadline: May 10th Looking forward to the discussions!!!
@rose_e_wang
Rose
22 days
How will ed tech change w LLMs? What is and isn't possible? If these Qs have been on your mind, submit your work to a workshop I'm organizing: Leveraging LLMs for Next Gen Ed Tech @ EDM 2024 by May 10th! ➡️ #EDM #EdTech
2
5
28
3
6
26
@rose_e_wang
Rose
8 months
I am thrilled to present at Women in Data Science! Come learn more about my work with @ddemszky on NLP and Education 😄 Topic: "Beyond Right or Wrong: Leveraging Language Models to Enhance the Learning Process" Aug. 30, 11am-12pm PDT Register now:
Tweet media one
1
2
25
@rose_e_wang
Rose
1 year
@ccanonne_ This is neat! Curious about what students wrote. One thing that’s been on my mind is how this might change grading rubrics to be less about a student’s generative ability (eg pseudocode, complexity analysis), and more about their discriminative ability (eg debug, critique).
1
1
20
@rose_e_wang
Rose
2 months
Let's build a future where every question leads to deeper understanding…for both students and teachers! 💡 Big thanks to @arankomatsuzaki for featuring our Backtracing work!! 🙏
@arankomatsuzaki
Aran Komatsuzaki
2 months
Backtracing: Retrieving the Cause of the Query - Proposes a new task called backtracing where the goal is to retrieve the cause of the query from a corpus - Shows limitations in current retrieval methods for performing backtracing repo: abs:…
Tweet media one
3
45
227
2
3
19
@rose_e_wang
Rose
2 years
Hi friends!! I'll be back at MIT for re-graduation on Monday for a week!! Let me know if you want to meet 🙏 + hang 🥳 + grab coffee ☕️ + eat food 🍜!! 😄
1
0
18
@rose_e_wang
Rose
2 years
This was done in collaboration with the amazing @esindurmusnlp , Noah Goodman and @tatsu_hashimoto ! 🥳 Many many thanks to my labmates from CoCoLab for their incredible support and feedback: @mike_h_wu , @GabrielPoesia , Ali Malik, @AlexTamkin ! 😎 🧵[T/T]
0
2
17
@rose_e_wang
Rose
2 years
Our insight is to represent _coherent_ language as a _smooth_ latent trajectory. We turn to Brownian bridge stochastic processes (SP) as a model for smooth trajectories. [4/T]
Tweet media one
1
1
16
@rose_e_wang
Rose
2 years
We train our encoder with contrastive learning (CL). Why CL? Bc of (a) its striking performance in learning representations (SimCLR for images @tingchenai ) & (b) exciting work in applying it to structured SPs ( @BingbinL )! [5/T]
Tweet media one
2
0
15
@rose_e_wang
Rose
2 months
Understanding why users ask questions is key because it’s a natural source of *content feedback*. We establish a diverse benchmark for backtracing and show traditional retrieval systems miss the mark in retrieving the cause of queries 😬 📎:
Tweet media one
1
0
14
@rose_e_wang
Rose
2 years
When we want language models to generate long text, they often output meandering text. One potential reason behind this failure mode is the model’s inability to plan ahead or represent long-range text dynamics. [2/T]
Tweet media one
1
0
13
@rose_e_wang
Rose
4 months
You want to understand how students learn, but not sure where to get started? 🎥 Here's a 2-min demo video of Edu-ConvoKit! It walks through its GPT-powered, quantitative and qualitative analysis tools. 👋 Happy exploring with Edu-ConvoKit!
@rose_e_wang
Rose
4 months
There’s a *lot* of insight about how students learn in education data like classroom conversations, but I can tell you it's *painful* to process and analyze. Introducing *Edu-ConvoKit*, a pipeline that handles the pre-processing, annotation, and analysis for you! 🧵
Tweet media one
2
19
95
0
3
12
@rose_e_wang
Rose
4 years
Our environment is designed to be easily configurable and light-weight. It’s perfect for folks interested in multi-agent systems or in compositional tasks/environments!
Tweet media one
1
0
13
@rose_e_wang
Rose
10 months
Lectures are a learning experience for students & teachers. Students learn about the subject & teachers learn about refining their instruction. But, online student feedback is unstructured. How can teachers learn from it?
Tweet media one
1
2
13
@rose_e_wang
Rose
2 months
Have questions about the work? Let's trace back to the source...: 📎: 💻: I’ll also be in Malta for #EACL presenting this work, so come chat with me there too 😉☀️
1
1
12
@rose_e_wang
Rose
2 years
On long text generation settings, Time Control (TC) preserves the text structure both in terms of ordering (up to +40% better) and text length consistency (up to +17% better). Human evaluators also prefer TC's output 28.6% more than the baselines. [8/T]
Tweet media one
1
0
11
@rose_e_wang
Rose
2 years
Prior work has explored remedies for this failure mode by using planning-based methods or implicitly learning text dynamics. However, these methods manually specify the text dynamics or sacrifice quality in long-horizon generation. [3/T]
1
0
10
@rose_e_wang
Rose
3 years
Yes!!! Make sure to check out @jeffclune + @ruiwang2uiuc 's awesome work at our 2nd poster session at 12:55pm PT! 🥳 GatherTown link can be found on our workshop's ICLR site.
@jeffclune
Jeff Clune
3 years
Our work Enhanced POET is an invited poster at this (excellent!) workshop in case you want to come ask questions or ask about any of our work. Thanks to the organizers for putting together such a wonderful event!
3
6
43
0
4
11
@rose_e_wang
Rose
4 months
👉 Check Edu-ConvoKit out: It’s easy and instrumental to transform the way we conduct research for improving real student learning outcomes. Work done at @StanfordEd @stanfordnlp @StanfordAILab with my advisor @ddemszky
1
3
11
@rose_e_wang
Rose
7 days
Experienced teachers engage their students in critical thinking—whereas novice tutors and LLMs don’t: They frequently give away the answer. Our work focuses on how experienced teachers do and *think* about remediating student mistakes. 📎:
Tweet media one
1
0
11
@rose_e_wang
Rose
2 years
The intuition is simple: The bridge imposes that a positive triplet (eg. three in-order sentences on Boston) makes up a smooth trajectory. A negative triplet should not construct a smooth trajectory (switching middle sentences with one on New York). [6/T]
Tweet media one
1
0
10
@rose_e_wang
Rose
2 years
After training the encoder, we finetune GPT2 to decode from past context and the encoded latent plan. At inference, we generate a latent plan by sampling from the bridge and conditionally generate each sentence using the latent plan. [7/T]
Tweet media one
1
0
10
@rose_e_wang
Rose
2 years
Lately, I’ve been super excited about teacher-student settings & thinking about how we can enable machines to (one day) reliably interact & _teach_ humans! If you’re interested in this direction, let's chat at @cells2societies poster session Fri April 29 8:15-9:05am PT! [6/N]
1
1
10
@rose_e_wang
Rose
10 months
(belated) paper laptop stickers for #ACL2023 !!! made with #stablediffusion 😃
Tweet media one
1
0
9
@rose_e_wang
Rose
3 years
Join us at the @ResistanceAI NeurIPS workshop! Currently we have a fireside chat with the amazing @red_abebe , @timnitGebru , @Abebab , @tetisheri . @ResistanceAI aims to create an inclusive space, discuss critiques of how #AI currently concentrates power & formulate resistance tactics.
@ResistanceAI
Resistance AI
3 years
Our first panel of the day is starting shortly, at 9am PST! 💜
Tweet media one
3
5
20
0
0
9
@rose_e_wang
Rose
2 years
Paper: w/ the amazing @mike_h_wu & Noah Goodman I’m grateful to many for sharing their insight, including my co-authors, Xiaojin Zhu (through his machine teaching works), @jayelmnop @DorsaSadigh @EmmaBrunskill @andyshih_ @willieneis @polynoamial [N/N]
0
0
8
@rose_e_wang
Rose
1 year
Super cool work from @KaitlynZhou --- generating expressions of uncertainty is extremely important in supporting human decision-making (and human reasoning)! My favorite part of the paper is their typology of uncertainty expressions (Table 4) 😀
@KaitlynZhou
Kaitlyn Zhou
1 year
Thanks so much for sharing our work! Our paper also discusses additional risks and opportunities that come with integrating expressions of uncertainty into LMs. Read the paper here: w/ @jurafsky @tatsu_hashimoto @stanfordnlp
0
2
15
0
4
8
@rose_e_wang
Rose
2 years
We also show that TC doesn’t sacrifice short/mid-range language modeling performance! Eg. TC matches/outperforms task-specific models like infilling by language modeling (ILM) on text-infilling or local representation methods on discourse coherence. [9/T]
Tweet media one
Tweet media two
1
0
8
@rose_e_wang
Rose
4 months
I am excited about language tools for education at scale because we move away from an oversimplified view of learning measured by the standardized test score….and towards language measures of student thinking and pedagogy. Repo:
Tweet media one
1
1
8
@rose_e_wang
Rose
2 years
Stop by @cells2societies poster session tomorrow (Fri) 8:15-9:05am PT if you're interested in teacher-student settings or the problem of "assessing then teaching models". Always excited to meet new folks too! 🙏🙂
@rose_e_wang
Rose
2 years
How can we algorithmically figure out what our model doesn’t know, and then construct datasets to improve it? We tackle this question in “Know thy student: Interactive learning with Gaussian processes” at #ICLR2022 @cells2societies workshop. Paper: [1/N]
Tweet media one
1
23
192
0
3
8
@rose_e_wang
Rose
2 months
What makes backtracing so hard?! Let’s use our Lecture dataset based on MIT OCW lectures as an example. Challenge 1: The queries don’t explicitly label what they’re caused by in the lecture/source document
Tweet media one
1
0
8
@rose_e_wang
Rose
1 year
@janleike @BlancheMinerva @MaksimSTW It would be great if OpenAI could make a more public blog post (eg. something with as much visibility as ChatGPT on ), clarifying the misunderstandings and also the implications (eg. SL fine-tuning on a good dataset ~= RLHF, and not RLHF>SL)!
2
0
7
@rose_e_wang
Rose
4 months
Tremendous thanks goes to the @lateinteraction , @JimMalamut , @lucy3_li for their feedback---esp. @lateinteraction for all the amazing, punny name suggestions (unfortunately, they didn't make the cut in my Slack polls!)🥲 Thanks for @YannHicke for contributions to the toolkit!
1
0
7
@rose_e_wang
Rose
3 years
I’ll be presenting our work in person at EMNLP. This will be my first NLP & in-person conference, so I’d love to meet folks! Please don’t hesitate to reach out 😄 📜: 📺: 👩‍💻:
0
0
7
@rose_e_wang
Rose
2 months
Challenge 2: The source document is *super* long. Long text understanding is important for backtracing to find the cause, but hard to nail.
Tweet media one
1
0
7
@rose_e_wang
Rose
2 months
Come hang out with us if you’re at #EACL2024 !!!
@krisgligoric
Kristina Gligorić
2 months
Attending #EACL2024 and passionate about computational social science (CSS) or social applications of NLP? Join us tomorrow (Wednesday at 2 pm in the Carlson room) for an informal CSS Birds of a Feather gathering! Let's meet, chat, and share insights!
Tweet media one
2
14
47
0
0
7
@rose_e_wang
Rose
3 years
Co-organized with the great @FeryalMP , @khimya , @LouisKirschAI , Annie Xie, Adam White, and Doina Precup at @iclr_conf --- see you all then!!!
0
0
7
@rose_e_wang
Rose
2 months
And Challenge 3: Backtracing relies domain expertise to understand the source document, the queries and the *reasons* behind the student’s confusion.
Tweet media one
1
0
7
@rose_e_wang
Rose
7 months
I'm excited about building NLP algs/systems to empower educators & students and enhance social interactions! ✨🚀🧑‍🎓👩‍🏫 Echoing the views of @ddmeyer in education and the innovative NLP x data science approaches of @Diyi_Yang & @timalthoff in mental health
1
0
6
@rose_e_wang
Rose
7 months
@abeirami @emnlpmeeting When can we expect a response from the PCs? We sent an email on Sunday and still haven't heard back. Our soundness/excitement scores were 5/4, 4/4, 4/4... 🙏 @juanmiguelpino @hbouamor Totally understand it's a busy time, but having an ETA would be helpful for resubmission! 🙂
2
0
5
@rose_e_wang
Rose
2 months
And thank you to @_akhaliq for his thread on our work as well 😃
@_akhaliq
AK
2 months
Backtracing Retrieving the Cause of the Query Many online content portals allow users to ask questions to supplement their understanding (e.g., of lectures). While information retrieval (IR) systems may provide answers for such user queries, they do not directly assist
Tweet media one
2
16
110
1
0
6
@rose_e_wang
Rose
2 months
And finally, a huge shoutout to my collaborators from @stanfordnlp @StanfordEd , esp @lateinteraction ! Working with Omar has been an incredible experience: He brings boundless insight and energy to the table, and I’m so grateful to learn from this star in our community! 🌟
Tweet media one
2
0
6
@rose_e_wang
Rose
7 days
We use Cognitive Task Analysis & work with experienced 🧑‍🏫s to surface their internal decisions. Patterns emerge: 🧑‍🏫 infer the *error type* -> determine an *intention* -> pick a *remediation strategy*.
1
0
7
@rose_e_wang
Rose
7 months
We have a lot of other exciting ongoing projects along this direction, so stay tuned!!! 📜 👩‍💻 🖥️ w/ the amazing Ashley Zhang, Carly Robinson, Susanna Loeb, and @ddemszky
2
2
6
@rose_e_wang
Rose
7 days
Our work complements recent work on self-improving LLMs with internal decision-making: But, we focus on eliciting and leveraging thoughts/decisions from real, experienced humans—in this case, teachers!
@ericzelikman
Eric Zelikman ✈️ ICLR
2 months
Language models today are trained to reason either 1) generally, imitating online reasoning data or 2) narrowly, self-teaching on their own solutions to specific tasks Can LMs teach themselves to reason generally?🌟Introducing Quiet-STaR, self-teaching via internal monologue!🧵
35
189
978
1
0
6
@rose_e_wang
Rose
7 days
We contribute a *unique, real* dataset: It contains the *internal decisions* paired with the response of these experienced teachers. It also includes real tutoring conversation snippets between novice tutors and students, across 120 math topics. Dataset:
Tweet media one
1
0
6
@rose_e_wang
Rose
7 days
There are so many cool findings in this work—I could go on for ages, but tweets can only do so much… For now I’ll leave y’all on a cliffhanger and show how experienced teachers make *diverse* and *complex* decision paths, compared to LLMs. Isn’t this beautiful!?
Tweet media one
2
0
7
@rose_e_wang
Rose
2 years
Eg. in an offline reinforcement learning setting, the student must navigate to the goal (green). The teacher determines states (yellow) the student has explored and accomplishes this task. The teacher can then construct demonstrations from states (orange) the student fails. [3/N]
Tweet media one
1
0
5
@rose_e_wang
Rose
3 years
Get ready for our panel happening in 10 mins!!! with @celestekidd @MelMitchell1 @SchmidhuberAI Adam White and Satinder Singh!!! 😀
Tweet media one
@neverendingrl
Never Ending RL
3 years
Please join us for a roundtable panel discussion with Celeste Kidd ( @celestekidd ), Satinder Singh, Melanie Mitchell ( @MelMitchell1 ), and Jürgen Schmidhuber ( @SchmidhuberAI ) moderated by Adam White. We are excited to hear this discussion. Don’t miss it!
Tweet media one
0
2
12
0
1
5
@rose_e_wang
Rose
4 months
@aryaman2020 @ChengleiSi Meanwhile @chengmyra1 , @krisgligoric and me *also* in the office 👀👀👀
1
0
5
@rose_e_wang
Rose
2 years
Make sure to check out @dharakyu & @jayelmnop 's cool work tomorrow (Friday) @EmeComWorkshop #ICLR2022 ! 🥳
@dharakyu
Dhara Yu
2 years
Excited to present “Emergent Covert Signaling in Adversarial Reference Games” at #ICLR2022 @EmeComWorkshop ! Paper: A thread: (1/n)
1
7
18
0
0
4
@rose_e_wang
Rose
1 year
tfw you pay for chatgpt+ but only get text-davinci-002 😢
Tweet media one
2
0
3
@rose_e_wang
Rose
3 years
Looking for a library AND viz tool for making ML training more efficient? Want to understand the trade offs you would make between cost, time and model quality? Check out what my cool friends @AveryLamp @abhi_venigalla @moinnadeem at @mosaicML recently released!!! 🥳😎
@DbrxMosaicAI
Databricks Mosaic Research
3 years
Hello World! Today we come out of stealth to make ML training more efficient with a mosaic of methods that modify training to improve speed, reduce cost, and boost quality. Read our founders' blog by @NaveenGRao @hanlintang @mcarbin @jefrankle (1/4)
Tweet media one
7
41
164
0
0
4
@rose_e_wang
Rose
28 days
@aryaman2020 @ChengleiSi This is so wholesome ❤️
0
0
4
@rose_e_wang
Rose
7 days
This work couldn’t have happened without the amazing support of my interdisciplinary team (CS+Education+Economics), and teachers of course! @stanfordnlp @StanfordEd @StanfordAILab Ashley @Carly__Robinson @loeb_susanna @ddemszky
1
0
5
@rose_e_wang
Rose
7 days
Our Bridge framework allows for two cool questions to be answered: - Can we use a human expert’s internal decisions to *improve* LLM responses? - Can we prompt LLMs to make their own internal decisions and self-improve?
1
0
5
@rose_e_wang
Rose
7 days
And yes, we *can* also prompt (some) LLMs to make decisions based on expert decision-making frameworks to self improve.
Tweet media one
1
0
5
@rose_e_wang
Rose
6 months
@tiwa_eisape @tallinzen @vansteenkiste_s @feishaAI @mhtessler @MIT @GoogleAI @GoogleDeepMind Super cool work on logical reasoning & biases like ordering effects! One thing I've been thinking about for a while is its connection to the task design lit in math ed (eg ).
Tweet media one
2
0
4
@rose_e_wang
Rose
7 days
What do we find? Yes, we *can* blend human expert decisions with LLMs and improve. (But still humans > LLMs with expert decision-making > LLMs)
Tweet media one
1
0
5
@rose_e_wang
Rose
2 years
We cast this problem as a teacher-student setup where the teacher must first interact to diagnose 🧪the student (the model), before teaching 👩‍🏫(constructing the training dataset). [2/N]
1
0
3
@rose_e_wang
Rose
1 year
@phillip_isola This is super cool!!! I've been exploring a similar idea for math education purposes and there are a lot of cool analogies we use 😀. It's interesting (and ominous??) that at age 30, we seem to hit a dark period though with our analogies
0
0
3
@rose_e_wang
Rose
3 years
@chelseabfinn @loseydp @DorsaSadigh @corl_conf Congrats everyone!!! 🥳 This is really amazing work!!! 😄
0
0
3
@rose_e_wang
Rose
7 months
Scaling high-quality tutoring is hard. With growing demand, many platforms use novice tutors who struggle to address mistakes and fail to seize learning opportunities. 📚But, turning mistakes into opportunities is key! Effective strategies can transform student understanding🚀
Tweet media one
1
0
3
@rose_e_wang
Rose
7 days
And finally, let’s not forget the amazing friends who gave feedback along the way :) ❤️ @cjziems @JoyHeYueya @GabrielPoesia @chengmyra1 @krisgligoric @malikrali @996roma @lateinteraction A lot of exciting, follow-up work cooking, so … stay tuned 👩‍🍳
1
0
6
@rose_e_wang
Rose
2 years
@ben_eysenbach Hey Ben! Neat work. Using contrastive learning for estimating value functions seems quite similar to the . They study it in symbolic domains + assume env simulator for negatives, so it's slightly different from your domains, but methodologically similar.🙂
1
0
3
@rose_e_wang
Rose
2 months
@AlexTamkin @MicahCarroll Also sharing this on tips for new researchers! Wrote this right after undergrad, so might need to update it with PhD/mentorship reflections 😁
0
0
3
@rose_e_wang
Rose
2 years
Note: The teacher can’t exhaustively probe the student esp. if (a) the state space is huge [this would take too long!], and (b) with limited communication [eg. if the student were a human, they would get exhausted!] [5/N]
1
0
3
@rose_e_wang
Rose
4 months
Let’s see Edu-ConvoKit in action! 👇 What are examples of real student reasoning in the classroom? Or, examples of how a tutor interacts with a student’s contributions? Boom, with a single function call with Edu-ConvoKit, we view these examples.
Tweet media one
Tweet media two
1
1
3
@rose_e_wang
Rose
3 years
Please consider submitting your work to our #ICLR2021 workshop! 😀
@neverendingrl
Never Ending RL
3 years
*** Deadline Extension Alert *** We have decided to extend our @iclr_conf workshop submission deadline by 48 hours, until February 28th AOE. Submit your work here: More details can be found on our website: #ICRL2021 #NERL2021
0
12
26
0
1
3
@rose_e_wang
Rose
4 years
Woops! I got nervous about my first tweet and tagged the wrong Max. Thanks @maxhkw for being an amazing research mentor! :)
0
0
3
@rose_e_wang
Rose
1 year
Big +1!! You learn how to do sanity checks, debug, and find edge cases through research.. Last summer, I heard an undergrad through my office wall bang on the table, then shout: "Ohhh myyy godddd!!!" *silence* "I'M CHANGING THE LIST AS I'M ITERATING THROUGH IT"😆 what a throwback
@AlexTamkin
Alex Tamkin 🦣
1 year
2) Learn to problem solve Google your errors / stack traces Play around with a library in an interactive python session (Maybe even ask a language model, as long as you can verify the answer it gives you!)
1
0
7
0
0
2
@rose_e_wang
Rose
7 months
Here’s my favorite qualitative example on a word problem: GPT4 gives away the solution, whereas the teacher suggests an illustrative strategy to help the student understand. Providing the student error and strategy label gets GPT4 to generate pedagogically better responses. 📈
Tweet media one
1
0
3
@rose_e_wang
Rose
3 years
And check out our multi-agent cooking environment here! 🍅🥬🍽️
@rose_e_wang
Rose
4 years
Excited to share gym-cooking, a *novel multi-agent Gym environment*: Based on recent work (, #CogSci2020 computational modeling prize winner) with amazing collaborators Sarah Wu, James Evans, Josh Tenenbaum, David Parkes, @maxkw !
6
60
305
0
0
3
@rose_e_wang
Rose
7 months
@ddmeyer @ddemszky On the other hand, I think this raises Qs abt how we should design collaborative systems/algs when users have very diff preferences in how they want to teach; this diversity Q comes up in other settings too e.g., mental health domain). Would love to chat & ideate about this!
1
0
1
@rose_e_wang
Rose
7 days
@cjziems @JoyHeYueya @GabrielPoesia @chengmyra1 @krisgligoric @malikrali @996roma @lateinteraction I also want to thank @ddmeyer for his insight & inspiring writing e.g., Teachers do so much behind the scenes. That is non-trivial to "chat-bot-away". But, figuring out where NLP can aid educators--that's the exciting challenge I want to focus on!
0
0
4
@rose_e_wang
Rose
7 months
Our work sheds light on the potential and current limitations of using LLMs to provide high-quality learning experiences for tutors & students at scale. We hope that it can serve as a valuable resource for providing equitable, high-quality learning experiences using *real* data.
1
0
3
@rose_e_wang
Rose
2 months
@ecekt2 @eaclmeeting @sandropezzelle @raquel_dmg This is really cool work!! Thanks for an awesome poster session 😃
1
0
3
@rose_e_wang
Rose
7 months
🔥Findings: Expert math teacher>LLMs+expert math teacher’s guidance>LLMs>novice tutors🔥 -The quality of the best LLM’s responses falls short of math teachers. -Providing LLMs w/ teacher’s strategy leads to a 75% improvement in response quality over models w/o that information.
Tweet media one
1
0
3