A great deal if you can get it indeed!
19.05.2025 15:06 β π 0 π 0 π¬ 1 π 0A great deal if you can get it indeed!
19.05.2025 15:06 β π 0 π 0 π¬ 1 π 0How?
19.05.2025 14:38 β π 0 π 0 π¬ 1 π 0
Could this be βReconstructing Training Data from Trained
Neural Networksβ?
giladude1.github.io/reconstructi...
proceedings.neurips.cc/paper_files/...
Am open-source alternative to (say) Slack or Discord: zulip.com
06.03.2025 07:20 β π 1 π 0 π¬ 0 π 0
It's a time management method
en.m.wikipedia.org/wiki/Pomodor...
Paper: openreview.net/pdf/a9c812c0...
Code: github.com/Flossiee/Hon...
ππΌπ»π²πππππ
- Introduces πππππππ, a dataset with 930 queries in six categories to evaluate LLM honesty
- Proposes curiosity-driven prompting and two-stage fine-tuning for improving honesty and helpfulness
- Demonstrates up to 124.7% honesty and helpfulness improvement in models like Mistral-7b
Figure 1: Fine-grained feedback from multimodal large language model help to yield more human-preferred images. Left: Output generated by the baseline text-to-image generative model. Right: Output generated by the baseline model optimized with fine-grained feedback from multimodal large language model. We illustrate improvements in generation quality across four aspects: PromptFollowing, Aesthetic, Fidelity and Harmlessness. See in Appendix for more visualization examples.
Multimodal Large Language Models Make Text-to-Image Generative Models Align Better
- VisionPrefer datset captures diverse preferences (prompt-following, aesthetic, fidelity, harmlessness) using multimodal LLMs
- VP-Score model matches human accuracy in preference prediction, guiding model tuning
Yeah, it would certainly be awesome to benchmark this empirically π
29.11.2024 08:02 β π 0 π 0 π¬ 0 π 0
It seems to be model dependent -- see for instance the GPT-3.5-Turbo vs. GPT-4 differences in here:
ar5iv.labs.arxiv.org/html/2310.18...
The Super Weight in Large Language Models
Setting as few as a single weight to zero will make various LLMs go from generating coherent text to outputting gibberish.
arxiv.org/abs/2411.07191
It unfortunately doesn't work that well with short (<200 tokens) responses.
www.nature.com/articles/s41...
Does the TULU paper count?
arxiv.org/abs/2306.04751
It doesn't need much more than a bit of gamification StackOverflow-style. Getting a bunch of badges for great reviews would go a long way.
Much of it seems to be low-hanging-fruit. E.g. my reviews were marked "Excellent" in the past but you cannot find it in my OpenReview.net profile.
Lobe Chat (github.com/lobehub/lobe...) + Ollama is a solid option
15.11.2024 16:09 β π 2 π 0 π¬ 0 π 0