github.com/albertz/wiki... Why game development is a great learning playground. Updated and resurrected article.
25.06.2025 21:10 β π 0 π 0 π¬ 0 π 0@albertzeyer.bsky.social
Deep Learning, speech recognition, language modeling, https://scholar.google.com/citations?user=qrh5CBEAAAAJ&hl=en Open source, https://github.com/albertz/
github.com/albertz/wiki... Why game development is a great learning playground. Updated and resurrected article.
25.06.2025 21:10 β π 0 π 0 π¬ 0 π 0github.com/albertz/py_b... www.reddit.com/r/Python/com... Some updates to my #Python better_exchook, semi-intelligently print variables in stack traces. Better selection of what variables to print, multi-line statements in stack trace output, full function qualified name (not just co_name)
19.06.2025 09:03 β π 0 π 0 π¬ 0 π 0I also see a bit the frustration in continuing doing research with small models and questioning the relevance of that nowadays. Or not being able to train large models with limited compute.
21.12.2024 22:58 β π 3 π 0 π¬ 1 π 0But I do see a bit the frustration that some of the domain-specific knowledge (e.g. subtleties of speech recognition models) seem to become somewhat irrelevant. But that was always clear. I'm not sure whether large models make this more irrelevant than what was anyway to be expected.
21.12.2024 22:58 β π 3 π 0 π¬ 1 π 0So I think working for 5 years or so full-time on this, involving research, publications in top conferences, provides still a much better level of experience than any bachelor or master student could possibly gain. Bachelor or master students are not better versed at training and deploying models.
21.12.2024 22:58 β π 3 π 0 π¬ 1 π 0At least in our group, the work we did was still very practical - always building actual speech recognition or translation or other kind of models, i.e. building things which could actually be used like this in production (and it is).
21.12.2024 22:58 β π 1 π 0 π¬ 1 π 0Google Scholar is messed up right now? The Transformer paper PDF links to some weird host, doesn't show other versions, and only shows the first author as sole author?
The same also for the LSTM paper after you click on 'cite'.
I often split this already in the beginning, so the loop goes like:
for msg_type, msg_opts in msg_queue: ...
msg_type is just a string, msg_opts is a dict.
Whether you use `match`/`case` or `if` on the msg_type is a matter of preference. Match/case is still quite young (Python 3.10).
I just learned that Torch ctc_loss calculates the wrong gradient (but when there was log_softmax before, it does not matter).
For the grad ctc_loss w.r.t. log_probs, it calculates exp(log_probs) - y, but correct would be -y. Some workaround: github.com/pytorch/pyto...
PS: First Bluesky post.