RecNet
Debates on the nature of artificial general intelligenceMelanie Mitchell
2024
Read
Yoav Artzi recommended on on 4/2/2024
Nothing new here, just a concise critique of the noise and hype around AGI, even if a little soft and gentle :)
Recent Contributions to The Mathematical Theory of CommunicationWarren Weaver
1949
Read
Yoav Artzi recommended on on 3/5/2024
A terrific read with a lot of relevant food for thought for current problems and challenges. It's also just an accessible way to get to know Shannon's theory (it was the introduction to Shannon's the Mathematical Theory of Communication).
LESS: Selecting Influential Data for Targeted Instruction TuningMengzhou Xia, Sadhika Malladi, Suchin Gururangan, Sanjeev Arora, Danqi Chen
2024
Read
Yoav Artzi recommended on on 2/20/2024
Cool technique for data selection with the goal of getting higher performance from fine tuning. Not sure I get the interpretability arguments, but a lot of excellent technical choices, and much to learn from.
On the Machine Learning of Ethical Judgments from Natural LanguageZeerak Talat, Hagen Blix, Josef Valvoda, Maya Indira Ganesh, Ryan Cotterell, Adina Williams
2022
Read
Yoav Artzi recommended on on 1/16/2024
A critique of the DELPHI paper. A lo of interesting food for thought. Some points are developed well, some less. The aspect of liability, an immediate issue with Ml models, could have been developed further.
Contrastive Preference Learning: Learning from Human Feedback Without Rl Joey Hejna, Rafael Rafailov, Harshit Sikchi, Chelsea Finn, Scott Niekum, W. Bradley Knox, Dorsa Sadigh
2023
Read
Yoav Artzi recommended on on 1/2/2024
Extending the DPO trick to advantage-based preference model, which is shown to be more sound in a recent work by Knox. What makes this paper a treat is the practical discussion of regularization, etc.
The Truth is in There: Improving Reasoning in Language Models with Layer-Selective Rank ReductionPratyusha Sharma, Jordan T. Ash, Dipendra Misra
2023
Read
Yoav Artzi recommended on on 12/26/2023
A pretty interesting result showing that removing components (without explicit optimization) from internal weight matrices in LLMs improves performance on some end tasks. Interesting connection to LORA, and much remain to outline boundaries of result.
QASA: Advanced Question Answering on Scientific ArticlesYoonjoo Lee, Kyungjae Lee, Sunghyun Park, Dasol Hwang, Jaehyeon Kim, Hong-in Lee, Moontae Lee
2023
Read
Yoav Artzi recommended on on 12/12/2023
An interesting formulation of answer composition process over multiple long documents.
Kiki or Bouba? Sound Symbolism in Vision-and-Language ModelsMorris Alper, Hadar Averbuch-Elor
2023
Read
Yoav Artzi recommended on on 11/28/2023
A really cool exploration of the presence of a fascinating linguistic phenomena within pre-trained models.
Meta-learning via Language Model In-context TuningYanda Chen, Ruiqi Zhong, Sheng Zha, George Karypis, He He
2022
Read
Yoav Artzi recommended on on 11/21/2023
A really nice and elegant connection between meta learning and in-context learning, solving some of the nested optimization difficulties of meta learning.
Fine-Tuning Language Models with Just Forward PassesSadhika Malladi, Tianyu Gao, Eshaan Nichani, Alex Damian, Jason D. Lee, Danqi Chen, Sanjeev Arora
2023
Read
Yoav Artzi recommended on on 11/14/2023
A very nice trick relying on predictable randomization that helps save memory in zero-order SGD. The focus in on LLMs, where reducing the memory needs has significant impact, but the method is pretty general.
© 2024 RecNet. All rights reserved.