Marek Suppa's Avatar

Marek Suppa

@mrshu.bsky.social

https://mareksuppa.com

31 Followers  |  406 Following  |  16 Posts  |  Joined: 23.08.2023
Posts Following

Posts by Marek Suppa (@mrshu.bsky.social)

A great deal if you can get it indeed!

19.05.2025 15:06 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

How?

19.05.2025 14:38 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Could this be β€œReconstructing Training Data from Trained
Neural Networks”?

giladude1.github.io/reconstructi...

proceedings.neurips.cc/paper_files/...

07.03.2025 16:20 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Zulip β€” organized team chat Zulip is an organized team chat app for distributed teams of all sizes.

Am open-source alternative to (say) Slack or Discord: zulip.com

06.03.2025 07:20 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Pomodoro Technique - Wikipedia

It's a time management method

en.m.wikipedia.org/wiki/Pomodor...

11.01.2025 16:40 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Paper: openreview.net/pdf/a9c812c0...

Code: github.com/Flossiee/Hon...

06.12.2024 21:07 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

π—›π—Όπ—»π—²π˜€π˜π—Ÿπ—Ÿπ— 

- Introduces π™ƒπ™Šπ™‰π™€π™Žπ™€π™, a dataset with 930 queries in six categories to evaluate LLM honesty

- Proposes curiosity-driven prompting and two-stage fine-tuning for improving honesty and helpfulness

- Demonstrates up to 124.7% honesty and helpfulness improvement in models like Mistral-7b

06.12.2024 21:06 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Paper: openreview.net/pdf?id=IRXyP...

05.12.2024 22:30 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Figure 1: Fine-grained feedback from multimodal large language model help to yield more
human-preferred images. Left: Output generated by the baseline text-to-image generative model.
Right: Output generated by the baseline model optimized with fine-grained feedback from multimodal
large language model. We illustrate improvements in generation quality across four aspects: PromptFollowing, Aesthetic, Fidelity and Harmlessness. See in Appendix for more visualization examples.

Figure 1: Fine-grained feedback from multimodal large language model help to yield more human-preferred images. Left: Output generated by the baseline text-to-image generative model. Right: Output generated by the baseline model optimized with fine-grained feedback from multimodal large language model. We illustrate improvements in generation quality across four aspects: PromptFollowing, Aesthetic, Fidelity and Harmlessness. See in Appendix for more visualization examples.

Multimodal Large Language Models Make Text-to-Image Generative Models Align Better

- VisionPrefer datset captures diverse preferences (prompt-following, aesthetic, fidelity, harmlessness) using multimodal LLMs

- VP-Score model matches human accuracy in preference prediction, guiding model tuning

05.12.2024 22:28 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Yeah, it would certainly be awesome to benchmark this empirically πŸ™‚

29.11.2024 08:02 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Large language models for aspect-based sentiment analysis Large language models (LLMs) offer unprecedented text completion capabilities. As general models, they can fulfill a wide range of roles, including those of more specialized models. We assess the perf...

It seems to be model dependent -- see for instance the GPT-3.5-Turbo vs. GPT-4 differences in here:

ar5iv.labs.arxiv.org/html/2310.18...

28.11.2024 19:16 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
The Super Weight in Large Language Models Recent works have shown a surprising result: a small fraction of Large Language Model (LLM) parameter outliers are disproportionately important to the quality of the model. LLMs contain billions of pa...

The Super Weight in Large Language Models

Setting as few as a single weight to zero will make various LLMs go from generating coherent text to outputting gibberish.

arxiv.org/abs/2411.07191

28.11.2024 09:15 β€” πŸ‘ 12    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0
Post image

It unfortunately doesn't work that well with short (<200 tokens) responses.

www.nature.com/articles/s41...

24.11.2024 09:47 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

Does the TULU paper count?

arxiv.org/abs/2306.04751

23.11.2024 21:38 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
OpenReview

It doesn't need much more than a bit of gamification StackOverflow-style. Getting a bunch of badges for great reviews would go a long way.

Much of it seems to be low-hanging-fruit. E.g. my reviews were marked "Excellent" in the past but you cannot find it in my OpenReview.net profile.

23.11.2024 19:35 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Lobe Chat (github.com/lobehub/lobe...) + Ollama is a solid option

15.11.2024 16:09 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0