Alexis Ross Profile Banner
Alexis Ross Profile
Alexis Ross

@alexisjross

2,633
Followers
898
Following
26
Media
368
Statuses

phd-ing @MIT_CSAIL , working on machine teaching | formerly nlp @allen_ai , comp sci & philosophy @harvard โ€˜20

Seattle, WA
Joined June 2018
Don't wanna be here? Send us removal request.
Pinned Tweet
@alexisjross
Alexis Ross
9 days
Good teachers *adapt* to student beliefs & misconceptions: Can LLM teachers? In new work w/ @jacobandreas , we introduce 1) the AdapT ๐Ÿ‘ฉโ€๐Ÿซ (Adaptive Teaching) evaluation framework & 2) AToM โš›๏ธ (Adaptive Teaching tOwards Misconceptions), a new probabilistic teaching method. (1/n)
Tweet media one
3
39
180
@alexisjross
Alexis Ross
2 years
Life update: Iโ€™ll be starting my PhD at @MITEECS & @MIT_CSAIL in the fall! Super excited to work with @jacobandreas , Yoon Kim, and the rest of the rich language ecosystem at MIT โœจ
46
18
521
@alexisjross
Alexis Ross
11 months
GPT-4 is already becoming widely-used as a writing assistant, but can it edit a scientific paper in response to reviews? ๐Ÿ“ In work led by Mike Dโ€™Arcy, we study this question & release ARIES, a dataset of paper edits aligned to specific reviewer comments:
Tweet media one
5
78
343
@alexisjross
Alexis Ross
3 years
Excited to share our preprint, "Explaining NLP Models via Minimal Contrastive Editing (MiCE)" ๐Ÿญ This is joint work with @anmarasovic and @mattthemathman Link to paper: Thread below ๐Ÿ‘‡ 1/6
5
45
227
@alexisjross
Alexis Ross
2 years
Sunset cruise to start off the PhD at @MIT_CSAIL โ˜€๏ธ Grateful to @MITEECS โ€™s GW6 for organizing!
Tweet media one
Tweet media two
3
8
211
@alexisjross
Alexis Ross
4 years
Excited to have won a Hoopes prize for my Computer Science/Philosophy thesis in explainable ML! Working on this interdisciplinary project was challenging but deeply rewarding. Forever grateful to my advisors Hima Lakkaraju & Bernhard Nickel for their invaluable guidance!
@hima_lakkaraju
๐™ท๐š’๐š–๐šŠ ๐™ป๐šŠ๐š”๐š”๐šŠ๐š›๐šŠ๐š“๐šž
4 years
Some good news: Alexis Ross who is one of my first undergrad thesis advisees at Harvard won the Hoopes Prize for her thesis. Yayy!
4
2
76
7
11
120
@alexisjross
Alexis Ross
3 years
Our upcoming #nlphighlights episodes will be a series on PhD applications. If there are any questions or topics you would like to see discussed, feel free to reply or send me a DM! We will look at these responses as we prepare our episodes ๐Ÿ˜„
11
14
119
@alexisjross
Alexis Ross
3 years
I'm happy to share that our paper "Explaining NLP Models via Minimal Contrastive Editing (MiCE)" was accepted into Findings of ACL 2021! Updated paper: Code & models: Work with @anmarasovic @mattthemathman
@alexisjross
Alexis Ross
3 years
Excited to share our preprint, "Explaining NLP Models via Minimal Contrastive Editing (MiCE)" ๐Ÿญ This is joint work with @anmarasovic and @mattthemathman Link to paper: Thread below ๐Ÿ‘‡ 1/6
5
45
227
3
18
103
@alexisjross
Alexis Ross
3 years
#nlphighlights 123: Robin Jia tells us about robustness in NLP: what it means for a system to be robust, how to evaluate it, why it matters, and how to build robust NLP systems. Thanks @robinomial and @pdasigi for a great discussion!
3
18
100
@alexisjross
Alexis Ross
9 months
*โƒฃ Resource alert for people applying to CS PhD programs this cycle *โƒฃ contains >60 example statements of purpose! It's made possible by the many generous submissions from new applicants, and new ones are always welcome! ๐Ÿ˜Š
@cs_sop_org
cs-sop.org
9 months
Are you thinking about your statement of purpose for grad school applications? โœ๏ธ If you are looking for good SoPs for reference, check out our ! Currently >60 SoPs ๐Ÿ“œ have been generously contributed by students from >25 institutions, across >15 CS fields.
1
14
61
1
25
91
@alexisjross
Alexis Ross
1 year
we heard about @stanfordnlp 's Alpaca and thought we should join in on the fun ๐Ÿฆ™ @gabe_grand @belindazli @zhaofeng_wu
Tweet media one
Tweet media two
2
5
88
@alexisjross
Alexis Ross
3 years
Happy to share that our paper, "Learning Models for Actionable Recourse," will appear in NeurIPS 2021! Very grateful to my collaborators/mentors @hima_lakkaraju and @obastani . Camera ready version coming soon!
@hima_lakkaraju
๐™ท๐š’๐š–๐šŠ ๐™ป๐šŠ๐š”๐š”๐šŠ๐š›๐šŠ๐š“๐šž
3 years
Learning Models for Actionable Recourse with @alexisjross and @obastani [5/n]
2
3
16
5
7
85
@alexisjross
Alexis Ross
2 years
Does training models with free-text rationales facilitate learning *for the right reasons*? ๐Ÿค” We ask this question in our #EMNLP2022 paper, "Does Self-Rationalization Improve Robustness to Spurious Correlations?" W/ @anmarasovic @mattthemathman ๐Ÿงต 1/n
2
19
80
@alexisjross
Alexis Ross
2 years
Happy to share that Tailor๐Ÿชก will appear at #ACL2022 as an oral presentation! For details, w/ new & improved results, check out our... - in-person talk (5/23, session 3) & poster (5/24, session 5) ๐Ÿ‡ฎ๐Ÿ‡ช - updated paper ๐Ÿ“ฐ: - code ๐Ÿ‘ฉโ€๐Ÿ’ป:
@tongshuangwu
Sherry Tongshuang Wu
3 years
New preprint alert! *Tailor: Generating and Perturbing Text with Semantic Controls* Title says it all: we perturb sentences in semantically controlled ways like how a tailor changes clothes ๐Ÿชก. w/ @alexisjross , @haopeng01 , @mattthemathman , @nlpmattg 1/n
Tweet media one
2
43
183
2
10
76
@alexisjross
Alexis Ross
1 year
Feeling grateful to have attended a wonderful #EMNLP2022 ! Highlights include the many interesting poster sessions and a memorable desert sunset ๐ŸŒ… Big thank you to everyone who stopped by our poster yesterday (and @i_beltagy for the ๐Ÿ“ธ)!
Tweet media one
Tweet media two
1
3
72
@alexisjross
Alexis Ross
2 years
Iโ€™ll be a mentor for MIT EECSโ€™s Graduate Application Assistance Program this application cycleโ€”Please consider signing up if youโ€™re applying to PhD programs this fall!
@_k_a_c_h_
Kartik Chandra (also on Mastodon and Bsky)
2 years
Applying to grad school in EE/CS this fall? โ€ฆneed help? Ask the MIT EECS Graduate Application Assistance Program! GAAP pairs applicants who need help with current PhD students for 1:1 mentoring. We match mentors weekly so it's never too late to sign up!
Tweet media one
1
40
123
3
7
67
@alexisjross
Alexis Ross
3 years
#nlphighlights 134: The 2nd episode in our PhD app series is on PhDs in Europe vs the US. @barbara_plank & Gonรงalo Correia share faculty & student perspectives on things to consider when choosing. We also discuss the ELLIS program. Cohosted w/ @zhaofeng_wu
3
15
64
@alexisjross
Alexis Ross
3 years
Very grateful to have attended #EMNLP2021 in Punta Cana! It was wonderful meeting so many virtually familiar and new faces in real life and discussing all things NLP (especially on the beach!) ๐Ÿ˜Š๐Ÿ
Tweet media one
Tweet media two
Tweet media three
1
0
64
@alexisjross
Alexis Ross
3 months
One of my favorite things about grad school has been getting to play chamber music again--Had a blast playing Tchaikovsky Piano Trio with @vikramsundar and @erencshin ๐Ÿ˜Š
1
1
51
@alexisjross
Alexis Ross
11 months
Our new preprint, led by @zhaofeng_wu , shows that traditional benchmark evals may over-estimate the generalizability of LLMs' task abilities ๐Ÿšจ We find LLM performance consistently drops on counterfactual variants of tasks (ex: code exec w/ 1-based indexing)! Details below ๐Ÿ‘‡
@zhaofeng_wu
Zhaofeng Wu
11 months
Language models show impressive performance on a wide variety of tasks, but are they overfitting to evaluation instances and specific task instantiations seen in their pretraining? How much of this performance represents general task/reasoning abilities? 1/4
Tweet media one
9
108
466
0
6
50
@alexisjross
Alexis Ross
10 months
Had a wonderful time at #DISI2023 over the past few weeks learning about diverse intelligences and exploring Scotland! ๐Ÿด๓ ง๓ ข๓ ณ๓ ฃ๓ ด๓ ฟ Grateful to be leaving with many new friends ๐Ÿ’™ @DivIntelligence
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
1
47
@alexisjross
Alexis Ross
2 years
The deadline for Predoctoral Young Investigator (PYI) applications for @ai2_allennlp is 2/15 โ€” Two days left to apply! I *highly* recommend the program for anyone interested in pursuing a PhD in natural language processing.
1
6
47
@alexisjross
Alexis Ross
2 years
#NeurIPS2021 Paper ๐Ÿ“ข: "Learning Models for Actionable Recourse" w/ @hima_lakkaraju & @obastani We'll be presenting this work at Poster Session 1. Happening tomorrow, Tues 12/7, 8:30-10 AM (PST). Come say hi! ๐Ÿ‘‹ Paper: More info:
Tweet media one
0
5
44
@alexisjross
Alexis Ross
1 year
In Abu Dhabi for #emnlp2022 ! Presenting a poster for our work on self-rationalization & robustness on Sunday at 11 AM: Iโ€™d love to chat about pragmatics, pedagogy, the relationship b/w explanations & learning, or anything in betweenโ€”please reach out! ๐Ÿค—
@alexisjross
Alexis Ross
2 years
Does training models with free-text rationales facilitate learning *for the right reasons*? ๐Ÿค” We ask this question in our #EMNLP2022 paper, "Does Self-Rationalization Improve Robustness to Spurious Correlations?" W/ @anmarasovic @mattthemathman ๐Ÿงต 1/n
2
19
80
1
2
40
@alexisjross
Alexis Ross
4 years
An interesting study on a real-world use of GPT-2: Generated โ€œdeepfakeโ€ comments were submitted to a federal public comment site for an Idhao Medicaid waiver and found to be indistinguishable from human comments. Paper: (1/2)
1
18
39
@alexisjross
Alexis Ross
5 years
Had a great time at #emnlp2019 presenting work with Ellie Pavlick, โ€œHow well do NLI models capture verb veridicality?โ€ ()Thank you Hong Kong and EMNLP for a great first conference!
Tweet media one
Tweet media two
Tweet media three
4
0
33
@alexisjross
Alexis Ross
3 years
Super excited to share our work on Tailor: a *semantically-controlled, application-agnostic system for generation and perturbation* and result of a really fun collaboration! Details in thread below๐Ÿ‘‡
@tongshuangwu
Sherry Tongshuang Wu
3 years
New preprint alert! *Tailor: Generating and Perturbing Text with Semantic Controls* Title says it all: we perturb sentences in semantically controlled ways like how a tailor changes clothes ๐Ÿชก. w/ @alexisjross , @haopeng01 , @mattthemathman , @nlpmattg 1/n
Tweet media one
2
43
183
0
4
29
@alexisjross
Alexis Ross
3 years
Ana has been an incredible mentor to me (and so many others), and I have no doubt she is going to make an equally incredible professor! Any institution would be very lucky to have her ๐Ÿ™Œ๐Ÿผ
@anmarasovic
Ana Marasoviฤ‡
3 years
Maybe this is also a good time to announce that I'm on the faculty job market โ€ผ๏ธ Reach out if Iโ€™m a good fit!
1
35
105
1
1
28
@alexisjross
Alexis Ross
1 year
Excited to share CREST, our new #ACL2023 work led by the awesome @MarcosTreviso ! This was a super fun collaboration w/ Marcos, @nunonmg , & @andre_t_martins ๐Ÿ˜Š CREST combines counterfactuals & rationales to improve model robustness / interpretability--details in the thread below๐Ÿ‘‡
@MarcosTreviso
Marcos Treviso
1 year
1/7 Thrilled to announce that our paper "CREST: A Joint Framework for Rationalization and Counterfactual Text Generation" has been accepted at #ACL2023 oral! ๐ŸŽ‰ This work is a result of a fantastic collaboration with @alexisjross , @nunonmg , and @andre_t_martins . Let's dive in!
1
7
39
0
2
27
@alexisjross
Alexis Ross
2 years
I am deeply grateful to my mentors, friends, & family, who helped me navigate all parts of the application process. Special thank you to everyone at @allen_ai for their support over the past 2 years ๐Ÿ’– I also feel so lucky to have met many wonderful people through this process!
1
0
26
@alexisjross
Alexis Ross
2 years
I spent two years as a predoctoral young investigator with @ai2_allennlp and could not have more positive things to say!! Please do apply if you want to work in an energizing and supportive environment with brilliant *and* kind people ๐Ÿ˜Š
@ai2_allennlp
AllenNLP
2 years
Prepare for a PhD program by doing a 1-3 year-long stint as a Predoctoral Young Investigator! Apply by tomorrow, 10/15:
0
2
19
0
1
23
@alexisjross
Alexis Ross
2 years
Go work with Ana!! ๐Ÿ‘‡๐Ÿป
@anmarasovic
Ana Marasoviฤ‡
2 years
I'm recruiting students! My interests include measuring usefulness of explanations for human-AI collaboration, addressing human factors that confound such measurements, & modeling interactive explainability (multimodality, few/zero-shot learning, dialogs, personalization, etc)
3
18
63
0
2
21
@alexisjross
Alexis Ross
4 years
#nlphighlights 121: Alona Fyshe tells us about the connection between NLP representations and brain activity in this episode hosted with Matt Gardner. Thank you @alonamarie and @mattg for a really interesting discussion on language and the brain!
2
6
18
@alexisjross
Alexis Ross
1 year
One of my favorite posters was this really cool work by @yuntiandeng @volokuleshov @srush_nlp (presented by @jxmnop ) on evaluating long-form generated text in the latent space
Tweet media one
0
2
18
@alexisjross
Alexis Ross
1 year
Check out my labmates cool work!
@akyurekekin
Ekin Akyรผrek
1 year
I am on the front page of MIT today! I am grateful to MIT News for covering my research! You can read the full paperย  I take the opportunity to support the people who suffered from the *unprecedented* earthquake in Turkiye. Trustworthy orgs to donate:
11
52
412
0
0
11
@alexisjross
Alexis Ross
3 years
Had a great time talking with @pdasigi and @thePetrMarek about the winning submission, Alquist 4.0, and how it can conduct coherent and engaging conversations! (Teaser: Alquist is designed to store and follow-up about personal details you mention, like that you have a brother)
@pdasigi
Pradeep Dasigi
3 years
#nlphighlights 132: @alexisjross and I chatted with Petr Marek @thePetrMarek about the Alexa Prize Socialbot Challenge, and this year's winning submission from Petr and team from the Czech Technical University. Thanks for the informative discussion, Petr!
0
2
6
0
1
10
@alexisjross
Alexis Ross
3 years
@emilypahn I also struggle with this! Whatโ€™s helped me is reading to write high level notes about the paperโ€™s main contributions, my takeaways/questions, and connections to what Iโ€™m working on. Writing answers to these qs as I read helps to focus my attention and know when to move on
0
0
10
@alexisjross
Alexis Ross
3 years
Big thank you to @complingy for the idea for this series and to everyone who sent me questions they wanted to see discussed. More topics will be covered in upcoming episodes ๐Ÿ™‚
0
0
9
@alexisjross
Alexis Ross
9 days
Lastly, a big thank you to my advisor @jacobandreas for being so supportive and making my first PhD project such a rewarding & fun experience ๐Ÿ˜Š (n/n)
0
0
8
@alexisjross
Alexis Ross
2 years
Overall, the variability of our results suggests that, despite the appeal of self-rationalization models for increasing model trustworthiness, self-rationalization training can have the unintended effect of *increasing* reliance on spurious features and biases ๐Ÿšจ 5/5
0
1
7
@alexisjross
Alexis Ross
3 years
@TheRealEGS I really like @bnickel34 โ€˜s advice on reverse outlining
1
0
5
@alexisjross
Alexis Ross
6 months
@wangzjeff @shannonzshen has built a cool slack chatbot :)
1
0
4
@alexisjross
Alexis Ross
3 years
@anmarasovic This looks so good and now I am craving fresh Mediterranean seafood ๐Ÿ˜ฉ
1
0
5
@alexisjross
Alexis Ross
3 years
@tiancheng_hu @996roma @complingy @nsubramani23 Thanks @tiancheng_hu and apologies for the delayed response! We weren't discussing any specifically, but there are a few other predoc/residency programs in industry that I know of--Here's a list I found (though I haven't personally looked through each one)
2
0
5
@alexisjross
Alexis Ross
11 months
In order for LMs to be effective writing assistants, they need to be able to model the relationship b/w paper feedback & revisions. Our findings highlight limitations in LMs' abilities to do so (eg focusing on surface-level meaning rather than underlying intent of feedback)
1
0
5
@alexisjross
Alexis Ross
3 years
Human explanations are *contrastive*โ€“They explain why an event happened *instead of* another event (the contrast case). Making model explanations contrastive could thus make them more user-friendly/useful. However, this property has largely been ignored in interpretable NLP. 2/6
1
2
5
@alexisjross
Alexis Ross
1 year
@jxmnop currently listening
Tweet media one
1
0
5
@alexisjross
Alexis Ross
2 years
As a current PYI on AllenNLP, Iโ€™ve gained invaluable hands-on research experience and preparation for a PhD โ€” and all in an incredibly collaborative, friendly, & supportive environment ๐Ÿ˜Š Please feel free to reach out over DM or email with any questions!
0
0
4
@alexisjross
Alexis Ross
1 year
@jxmnop Now eagerly awaiting this feature to be rolled out to my account ๐Ÿฅฒ
1
0
4
@alexisjross
Alexis Ross
11 months
We hope ARIES will aid researchers in studying the paper revision process and developing new methods for assisting authors and reviewers in the writing process!
1
0
4
@alexisjross
Alexis Ross
5 months
@shannonzshen and like RAG, a cheatsheet is not enough to completely rule out hallucinations at test time ๐Ÿ˜…
0
0
4
@alexisjross
Alexis Ross
2 years
@kayo_yin @SCSatCMU Youโ€™re a star!!! โญ๏ธ
0
0
4
@alexisjross
Alexis Ross
3 years
Episodes will be uploaded here:
0
0
4
@alexisjross
Alexis Ross
3 years
MiCE has 2 stages: In Stage 1, we train an Editor model to make edits targeting given contrast labels. In Stage 2, we use the Editor to make edits using both binary search and beam search to find edits resulting in the highest contrast prediction probabilities from the model. 4/6
Tweet media one
1
0
4
@alexisjross
Alexis Ross
3 years
@LakeBrenden @glmurphy39 Looks very interesting, looking forward to reading! (5) reminds me of work by @belindazli @Maxwell_Nye @jacobandreas showing that word reps encode changing entity states based on inputs. Wonder if this would also hold for facts like "Dolphins are mammals"
0
0
4
@alexisjross
Alexis Ross
9 days
In AdapT, a teacher aims to teach a target concept to a student who has unknown misconceptions. AdapT includes 3 domains: 1) fraction arithmetic 2) English verb conjugation 3) function learning. Students are both simulated ๐Ÿค– & human ๐Ÿ‘ฉโ€๐Ÿ’ป. (3/n)
Tweet media one
1
0
5
@alexisjross
Alexis Ross
9 days
Using AdapT, we evaluate both GPT4 and probabilistic teaching methods. We introduce AToM โš›๏ธ, which performs online inference of student priors, then selects informative teaching examples based on these inferences. (4/n)
1
0
4
@alexisjross
Alexis Ross
3 years
We present Minimal Contrastive Editing, or MiCE, a two-stage approach to generating contrastive explanations of model predictions. A MiCE explanation is a modification of an input that causes the model being explained to change its prediction to a given contrast prediction. 3/6
Tweet media one
1
0
3
@alexisjross
Alexis Ross
3 years
Finally, we show how MiCE edits can be used for two use cases in NLP system developmentโ€“discovering dataset artifacts (ex: IMDB edit below) and debugging incorrect model predictions (ex: RACE edit below). Feel free to to reach out with any questions or comments! 6/6
Tweet media one
1
0
3
@alexisjross
Alexis Ross
4 years
More from the paper: Federal comment sites are not currently equipped to detect such automated submissions. The ease with which deepfake text can be created and used highlights the need for technological reforms (2/2)
0
0
3
@alexisjross
Alexis Ross
11 months
ARIES includes a synthetic training dataset of 3.9k examples and a manually-annotated test set of 196 examples, using computer science papers, reviews, and author responses from OpenReview.
1
0
2
@alexisjross
Alexis Ross
3 years
1
0
3
@alexisjross
Alexis Ross
2 years
@anmarasovic And soon afterwork skiing? โ›ท๏ธ
1
0
3
@alexisjross
Alexis Ross
4 years
@hima_lakkaraju Thank you so much for all of your indispensable advising and support!! ๐Ÿ˜Š
0
0
3
@alexisjross
Alexis Ross
2 years
We realize many interviews in this cycle have already happened, but we hope this episode is still useful for people currently navigating visit days/PhD decisions (and for future applicants)!
0
0
3
@alexisjross
Alexis Ross
9 days
We also find that AToM makes more *accurate inferences* about student beliefs (both simulated & human) than GPT4. It also selects *key teaching examples* (i.e. that target student misconceptions) earlier in teaching. (6/n)
Tweet media one
1
0
3
@alexisjross
Alexis Ross
11 months
In particular, GPT-4 likes to rigidly follow instructions & paraphrase the comment it is responding to, and it includes fewer technical details than real edits.
Tweet media one
1
1
3
@alexisjross
Alexis Ross
3 years
@lambdaviking @NYUDataScience This is the best phd announcement yet ๐Ÿคฃ
0
0
3
@alexisjross
Alexis Ross
2 years
We train 6 model types on NLI & commonsense QA with/without free-text rationales and measure robustness to spurious correlations through 1) challenge datasets 2) test sets where reliance on spurious correlations would lead to incorrect answers ๐Ÿ‘ฉโ€๐Ÿซ 2/n
Tweet media one
1
0
2
@alexisjross
Alexis Ross
3 years
@joeddav So awesome! We should have brought papers to read on our HK hike ๐Ÿ˜†
1
0
2
@alexisjross
Alexis Ross
3 days
@jwanglvy Great q! You're right; ATOM is a probabilistic method with main components: set of Bayesian student models & set of possible examples to choose from. It tracks student predictions and chooses both a student model & an optimal teaching example at each step. Hope that clarifies!
1
0
1
@alexisjross
Alexis Ross
3 days
Cool new work by @seungwookh and @IdanShenfeld on aligning/personalizing models at decoding time!
@seungwookh
Seungwook Han
3 days
๐Ÿš€ Stronger, simpler, and better! ๐Ÿš€ Introducing Value Augmented Sampling (VAS) - our new algorithm for LLM alignment and personalization that outperforms existing methods!
Tweet media one
5
31
123
0
0
4
@alexisjross
Alexis Ross
4 years
@delliott The assignments for this Harvard NLP class were to reproduce various papers and might be a good place to look:
0
0
2
@alexisjross
Alexis Ross
2 years
While results are model/task-specific, we observe some general trends ๐Ÿ“ˆ: - Data: Improvements tend to be in lower-resource settings & self-rationalization can hurt in higher-resource settings - Model size: Within model families, larger models benefit more from rationales 3/n
Tweet media one
1
0
2
@alexisjross
Alexis Ross
2 years
We also find that *rationale content* affects results. Training with positive rationales in the ECQA dataset improves robustness, while using freeflow/negative rationales harms robustness. 4/n
Tweet media one
1
0
2
@alexisjross
Alexis Ross
2 years
0
0
2
@alexisjross
Alexis Ross
9 days
Consider the 3rd-grader who answers 1/2 + 2/4 = 3/6 โŒ A good teacher might immediately develop a hypothesis about the studentโ€™s misconceptionsโ€“over-generalizing the rule for multiplication (adding both nums & denoms)โ€“which should influence their course of instruction. (2/n)
1
0
3
@alexisjross
Alexis Ross
1 year
0
0
2
@alexisjross
Alexis Ross
2 years
@aaronsteven You might also be interested in our work on Tailor (), which guides generation with control codes that derive from PropBank representations
1
0
2
@alexisjross
Alexis Ross
9 days
On the whole, our results point to complementary advantages of LLM-based teachers like GPT4 and more structured models like AToM. An interesting direction for future work would be to look into combining these to get the best of both worldsโ€“See our paper for more! (8/n)
1
0
2
@alexisjross
Alexis Ross
2 years
@tallinzen @RTomMcCoy @jacobandreas Congrats Tom!! ๐ŸŽ‰๐Ÿฅณ
1
0
2
@alexisjross
Alexis Ross
11 months
When tasked with writing edits given a reviewer comment (eg as a writing assistant), GPT-4 often produces content that looks reasonable on a surface level, but has systemic differences from human-written edits.
1
0
2
@alexisjross
Alexis Ross
3 years
Experiments on classification/multiple-choice Q&A show that MiCE edits are not only contrastive, but also *minimal* and *fluent*, consistent with human contrastive edits. 5/6
Tweet media one
1
0
2
@alexisjross
Alexis Ross
1 year
@yanaiela working on it ๐Ÿ˜„
0
0
2
@alexisjross
Alexis Ross
11 months
This indirectness is compounded by the fact that authors sometimes disagree with a request or get creative with how to address it.ย We find that GPT does worse at aligning indirect comments and non-compliant edits.
Tweet media one
2
0
2
@alexisjross
Alexis Ross
11 months
Identifying which comments correspond to which edits is hard, even for GPT-4.ย Many review comments are nuanced & indirect; instead of saying "You should use a more realistic dataset, like ImageNet", they might say "I am not convinced that the dataset used is realistic".
Tweet media one
1
0
2
@alexisjross
Alexis Ross
6 months
@neuranna Congratulations Anya!! ๐Ÿฅณ We will miss you at MIT!
0
0
2
@alexisjross
Alexis Ross
1 year
@joeddav @KahlertSoC @UUtah @UtahNLP Huge congrats Joe!! ๐Ÿฅณ So excited for you :)
1
0
2
@alexisjross
Alexis Ross
9 days
On the other hand, GPT4 seems to have encoded information harder to represent in structured methods like AToMโ€“for example, that it may be easier for human students to learn from examples closer to the origin when trying to learn the weights of a line. (7/n)
Tweet media one
1
0
2
@alexisjross
Alexis Ross
9 days
We find that simulated students ๐Ÿค– learn more efficiently with AToM than GPT4. In experiments with human students ๐Ÿ‘ฉโ€๐Ÿ’ป, both AToM & GPT4 outperform random example selection. (5/n)
1
0
3
@alexisjross
Alexis Ross
2 years
@nlpnoah Aw thank you so much Noah ๐Ÿ˜Š I am very grateful for your support!!
0
0
1