YouTube video by PyCon US
Building Scalable AI Tool Servers with Model Context Protocol (MCP) and Heroku (Sponsor: Heroku)
Here is a recording of my live demo at PyCon US 2025 on building scalable AI tool servers using the Model Context Protocol (MCP) and Heroku
www.youtube.com/watch?v=01I4...
29.05.2025 16:50 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
Elon Musk endorses far-right German political party, wading deeper into global politics | CNN Business
Musk, the billionaire Trump ally who is playing a public role in the incoming administration, posted in support Friday of Alternative for Germany, or AfD, after the German government collapsed this we...
I was surprised at how clear-cut and blatant it was. I mean, two times in a row, closed fingers, correct angle.
Meanwhile, Musk has recently issued public support for the far-right wing AfD party, often described as anti-semetic / extremist.
www.cnn.com/2024/12/20/m...
That + no apology...
22.01.2025 00:29 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
In honor of MLK day, here's super interesting essay my partner wrote on Martin Luther King Jr: what he actually believed and accomplished (different than what is sometimes described).
docs.google.com/document/d/1...
Incredibly impressive person.
21.01.2025 03:49 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
Nice! Would love to be added (11 yrs in AI, co-author of Malware Data Science, love them NNs)
07.01.2025 16:46 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Am I reading this right? Techniques to make the model safe again had almost no effect on non-small models :o.
03.12.2024 21:46 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
03.12.2024 21:46 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Sleeper Agents
arxiv.org/pdf/2401.05566
So many AI safety issues get worse, & harder to combat the larger and more advanced your model gets:
"The backdoor behavior is most persistent in the largest models and in models trained to produce chain-of-thought reasoning"
03.12.2024 21:46 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
03.12.2024 21:10 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
A response to X is going to be (usually) written by someone socially, politically near X's author, vs some other random piece of content Y.
It's extremely hard to take out sycophancy out of an LLM, trained the way we train them.
03.12.2024 21:00 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Anthropic's "Towards Sycophancy In Language Models" arxiv.org/pdf/2310.13548
TLDR: LLMs tend to generate sycophantic responses.
Human feedback & preference models encourage this behavior.
I also think this is just the nature of training on internet writing.... We write in social clusters:
03.12.2024 21:00 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Say a model learns strategy x to minimize training loss --> Later, min(test loss) involves strategy y, but the model regardless sticks with strat x (inner misalignment).
Assuming outer misalignment, x can be seen as safer than y.
That being said, the better the model, the less this will happen.
24.11.2024 17:01 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
In AI safety, we have inner misalignment (actions don't minimize the loss function) and outer misalignment (loss function is misspecified).
But I do think that inner misalignment (~learned features) tend to act as a protective mechanism to avoid outer misalignment implications.
I, er, really hope.
23.11.2024 15:46 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Princeton computer science prof. I write about the societal impact of AI, tech ethics, & social media platforms. https://www.cs.princeton.edu/~arvindn/
BOOK: AI Snake Oil. https://www.aisnakeoil.com/
Philosopher/AI Ethicist at Univ of Edinburgh, Director @technomoralfutures.bsky.social, co-Director @braiduk.bsky.social, author of Technology and the Virtues (2016) and The AI Mirror (2024). Views my own.
AI Security @ NVIDIA
OSS Security @ Project Jupyter and NumFOCUS
https://developer.nvidia.com/blog/author/jolucas/
DSc in Cybersecurity. Proud Apache. AI security researcher. Lover of books, cats, dogs, and movies. ARVA board member. AI Village officer. DEFCON goon.
Bass guitarist for the BCGS Orchestra, double bass with BGSO and VVGO, reenigne, ML/AI vivisectionist
Hackers, ML researchers, and data scientists focused on the use and abuse of AI; join us!
Discord: http://discord.gg/xMK7fuu
Twitch: https://twitch.tv/aivillage
Building recommender systems @ Consumer Tech Co
Leading Cybersecurity AI @NVIDIA | Cyber+ML/DL researcher | Engineering the future of cyber | views == mine
Climber ๐ช๐ฌ |Chief AI Scientist at @ainowinstitute.bsky.social | Safety engineer (nuclear, software & AI/ML).
|x- Trail of Bits, Zipline, OpenAI, Adelard, MSFTResearch
https://www.heidyk.com/
Supply chain security @ Google OSS Security Team. Previously TensorFlow Security & OSS (@ Google); Haskell+differential privacy+ML @ LeapYear.
I should be stretching
ml + privacy
https://github.com/jvmncs
busy building stuff. likes: offensive security, LLMs, and dumb memes. prev: research scientist @ OpenAI / CS PhD @ Harvard / cofounded DEF CON AI Village
๐ฆBetter software for better government ๐ฆ
goingforbrooke.com
traffickingfreetomorrow.com
Tweets about natural language processing, machine learning, infosec, and general trivia. Opinions my own.
Machine Learning @ Sublime Security
Language Modeling nerd. Lo-Fi Junkie.
MPLS
Senior Director of Research. Black Hat Review Board Member (AI, ML, and DS track lead) and International public speaker. I focus on emerging technologies and risks at the intersection of humanity and tech. Hype Critic. My writing: https://perilous.tech
tech, cybersecurity, but mostly politics these days
denver
GO BLUE
Working in the intersection of data science & cyber security / I know more about x509 and TLS than your average person out walking their cat / I collect hobbies like other people collect comic books. turbocodr on that other site
Member of Technical Staff @ Runway | ex-Wayve | Oxford PhD
Living in London