Done.
09.06.2025 15:11 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0@kfountou.bsky.social
Associate Professor at CS UWaterloo Machine Learning Lab: opallab.ca
Done.
09.06.2025 15:11 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0I wrote a blog post about it.
Link: medium.com/@kimon.fount...
2) Can a neural network discover instructions for performing multiplication itself?
The answer to the first question is yes, with high probability and up to some arbitrary, predetermined precision (see the quoted post).
Link to the paper: arxiv.org/abs/2502.16763
Link to the repository: github.com/opallab/bina...
Learning to execute arithmetic exactly, with high probability, can be quite expensive. In the plot, 'ensemble complexity' refers to the number of independently trained models required to achieve exact learning with high probability. ell is the number of bits per number in the input.
26.05.2025 03:21 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0New paper: Learning to Add, Multiply, and Execute Algorithmic Instructions Exactly with Neural Networks
26.05.2025 03:21 โ ๐ 5 ๐ 1 ๐ฌ 1 ๐ 1Learning to execute arithmetic exactly, with high probability, can be quite expensive. In the plot, 'ensemble complexity' refers to the number of independently trained models required to achieve exact learning with high probability. ell is the number of bits per number in the input.
26.05.2025 03:19 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0I never understood the point of trams. They're slow and expensive. I've been to two cities that built them while I was there, Edinburgh and Athens, and in both cases, the projects were born out of corruption. Especially in Edinburgh, it was a disaster. en.wikipedia.org/wiki/Edinbur...
21.05.2025 14:39 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0Thanks, the connection to formal languages is quite interesting. I have a section in the repo regarding formal languages but it's small mainly because it's not a topic that I am familiar with. I will add them!
17.05.2025 11:49 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Update, 14 empirical papers added!
16.05.2025 17:26 โ ๐ 3 ๐ 0 ๐ฌ 0 ๐ 0The SIAM Conference on Optimization 2026 will be in Edinburgh! I donโt really work on optimization anymore (at least not directly), but itโs cool to see a major optimization conference taking place where I did my PhD.
15.05.2025 13:32 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Currently NeurIPS has 21390 submissions. The final number last year was 15671.
Observation made by my student George Giapitzakis.
Got a pin this morning
Einstein problem: en.wikipedia.org/wiki/Einstei...
Our new work on scaling laws that includes compute, model size, and number of samples. The analysis involves an extremely fine-grained analysis of online sgd built up over the last 8 years of understanding sgd on simple toy models (tensors, single index models, multi index model)
05.05.2025 17:08 โ ๐ 5 ๐ 1 ๐ฌ 0 ๐ 0Hey, I definitely predicted this correctly.
01.05.2025 20:42 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0ChatGPT gives me the ability to expand my search capabilities on topics that I can only roughly describe, or even illustrate with a figure, when I donโt know the exact keywords to use in a Google search.
01.05.2025 17:58 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0That's a comprehensive study on the expressivity for parallel algorithms, their in- and out-of-distribution learnability, and it includes a lot of experiments.
link: arxiv.org/abs/2410.01686
Positional Attention is accepted at ICML 2025! Thanks to all co-authors for the hard work (64 pages). If youโd like to read the paper, check the quoted post.
01.05.2025 13:22 โ ๐ 7 ๐ 1 ๐ฌ 1 ๐ 0NeurIPS 2026 in the Cyclades. Just saying.
29.04.2025 18:12 โ ๐ 4 ๐ 0 ๐ฌ 0 ๐ 0Wait, isn't that America?
28.04.2025 17:29 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0This is different from simply replacing the discontinuous activation in the neural network with a continuous one and then using standard NTK.
link: openreview.net/forum?id=kfd...
They analyze a modified gradient flow, where the Jacobian for the training data uses an approximate derivative of the activation function.
27.04.2025 00:03 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0I enjoyed reading the paper "A Generalized Neural Tangent Kernel for Surrogate Gradient Learning" (Spotlight, NeurIPS 2024).
They extend the NTK framework to activation functions that have finitely many jumps.
โ๏ธ Code Shaping, an AI-powered software, allows users to edit their code through sketches like diagrams and graphs ๐
๐ This game-changing platform won the Best Paper Award at #CHI2025.
๐Read more: uwaterloo.ca/computer-sci...
#UWaterloo #AI
Regarding this particular case, I can read exactly what the code is doing and it seems quite interpretable. I don't have to know some other library to understand the code. It seems to only use basic instructions.
23.04.2025 22:00 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0I prefer more verbose coding, but again, I am only doing prototyping. I find it very annoying when someone's code includes efficient shortcuts that are hard to interpret without a lot of experience.
23.04.2025 22:00 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0To be honest, I find the code by the chat way more readable, exactly because it's verbose. But, I suck at coding, I only prototype ideas...
23.04.2025 16:30 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0With my first Bluesky post, I am very pleased to share that my last PhD paper "Deterministic complexity analysis of Hermitian eigenproblems" has been accepted in ICALP 2025. A preprint is available on Arxiv:
arxiv.org/abs/2410.21550
A bit more info on linkedin: www.linkedin.com/posts/aleksa...