Sign in to view Vincent’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Vincent’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
London, England, United Kingdom
Sign in to view Vincent’s full profile
Vincent can introduce you to 10+ people at Periodic Labs
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
13K followers
500+ connections
Sign in to view Vincent’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Vincent
Vincent can introduce you to 10+ people at Periodic Labs
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Vincent
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Vincent’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Websites
- Personal Website
-
https://github.com/vmoens
- Company Website
-
https://github.com/vmoens
About
Ph.D. in Computational and Cognitive…
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
Activity
13K followers
-
Vincent Moens shared thisHere's a side quest I've been working on for a little while: mujoco-torch (https://lnkd.in/eN82F9hP) I started this during my time at Meta in my spare time as a way to demonstrate torch.compile usage for physics simulation in #PyTorch. Since then, I worked during my weekends to lead it to a more mature state. It's now available on PyPI. `pip install mujoco-torch` Running physics in PyTorch (or JAX) instead of a traditional C simulator isn't just about speed -- it means your simulator lives in the same framework as your learning pipeline. Gradients flow through the physics, batching is a single torch.vmap call, and torch.compile fuses the entire simulation step into optimized GPU kernels with zero Python overhead. No serialization boundaries, no cross-framework data copies -- just tensors all the way down. ⚙️ What it does: mujoco-torch is a full PyTorch port of MuJoCo MJX -- forward dynamics, constraint solving (CG and Newton), frictional contacts, 30+ sensor types, and 12 collision functions covering all major geom types. It's numerically equivalent to MJX, verified at float64 precision. You can load (hopefully) any MuJoCo XML, move it to GPU, and simulate with a few lines of code. The end goal is E2E compilation of your simulation and training loop! (I've also added a small rendering util that alleviates all the headless EGL rendering nightmares you might have had with with MuJoCo!) 🔦 Demo: The repo also ships with an RL environment zoo built on TorchRL: HalfCheetah, Ant, Hopper, Walker2d, Swimmer, and Cartpole, all trainable with SAC or PPO out of the box. These environments served as end-to-end validation that the physics actually produces learnable behaviors. 🚀 Perf: with torch.compile + torch.vmap at large batch sizes, we see millions of simulation steps per second on a single GPU -- up to 3.6M steps/s on Half-Cheetah and 2.5M on Humanoid at batch size 32k. That's roughly 40x the throughput of MuJoCo C running sequentially on CPU. The library is designed for the regime where you're running thousands of parallel environments, which is exactly where modern RL training operates. Hope you'll enjoy it! #robotics #AI #ReinforcementLearning
-
Vincent Moens posted thisI’ll be attending NeurIPS in San Diego next week! Reach out if you want to meet
-
Vincent Moens shared thisAppreciate the shoutout and the creative read on who we are and what we do :) We actually have fewer ex-OpenAI folks than people seem to think — and more scientists than ML engineers. It turns out it takes a surprisingly large number of chemists and physicists to make an automated lab work. As for the whole ‘destined to be bought by Meta’ angle — that’s a fun one! It's interesting to see what the outside world imagines we’re up to. And on the focus bit: not sure I see why funding makes you lose focus but I appreciate the advice!Vincent Moens shared thisWhat to do when a competitor raises 💰$300M on a pitch deck Periodic Labs announced their mind-bending seed fundraise 1 month ago, here is what we learned over this month at Entalpic: 1️⃣ Check your ego. If A16z just bet big on your field, that’s actually a good thing. It means your domain is heating up : smart money sees potential 🔥. Take a breath and be proud you’re building in a space that matters. 2️⃣ Understand what’s really happening. The team is ex-OpenAI 🤖, super AI-driven, not chemistry or materials focused. From what I’ve heard, the crazy valuation reflects how expensive “acquihires” are in that world. Many US VCs told me they see a likely scenario: being bought by Meta or another giant within a couple of years for billions 💸. 3️⃣ Act accordingly. Big rounds can be both a blessing and a trap. With that much cash, it’s easy to go broad, market everywhere, lose focus. I’ve been there. The best move for smaller players? Focus 🎯. Pick one market, own it, and grow from there. Depth beats breadth every time. 4️⃣ Congratulate them, sincerely. So, kudos to Periodic Labs 👏 for this wild raise. Please, don’t underdeliver : our entire field needs you to succeed 🙏. Let’s push our science beyond skepticism and make this real. Good luck to all of us building the future of materials!
-
Vincent Moens shared thisVincent Moens shared thisUltralytics is heading to the PyTorch Conference in San Francisco! 🚀 Join us from October 22 - 23 at #PyTorch2025 as we showcase the latest advancements in Vision AI and demonstrate how Ultralytics YOLO models are driving innovation across industries. Come meet the team, explore live demos, and discover how Ultralytics is shaping the future of AI-powered vision! 📍 Booth S7 📅 October 22 - 23, 2025 | Moscone West, San Francisco
-
Vincent Moens shared thisWaymo coming to #London is a major step forward for the UK — and for Europe. Hopefully this will help unlock this tech for the rest of the continent too! Some healthy competition for the homegrown Wayve, I guess... Self-driving cars are much safer for both passengers and everyone around them, and they’re also better for the environment. As a cyclist (and an occasional rider), I couldn’t be more enthusiastic! There’s a funny paradox in the self-driving debate: Before the tech was mature, people used to ask, “What should a car do if it has to choose between swerving and killing its passengers or staying the course and hitting 10 people?” It turns out that question was largely irrelevant. In the US alone, if every car were a Waymo, we’d avoid around 35,000 deaths a year and save roughly $1 trillion in societal costs. The so-called “moral dilemma” barely registers — the tech is so much safer that situations like that become epsilon in the grand scheme of things. But... I really can’t wait to see how they handle yellow boxes though 👀 https://lnkd.in/e3NBNPxU Sources: https://lnkd.in/eVuMVhyz
-
Vincent Moens shared thisA great blend of PyTorch and kt for dynamically building powerful RL pipelines Highly recommend it!Vincent Moens shared this“Programmatic control over ML compute is the new world. Now that I've seen it I can never go back to Slurm. It's hard to point to just one thing that's better, it's 100 things.” Announcing Kubetorch, fixing ML and RL development on Kubernetes: https://lnkd.in/ewBa6CXd We're also excited to announce our $5M seed round led by Work-Bench, with Hetz Ventures, Fathom, Andrew Miklas (YC, Pager Duty), Spencer Kimball (Cockroach Labs), Sahir Azam (MongoDB), Lauryn Motamedi (Notion), Nico C. (Dust), and Romain Huet (OpenAI, Stripe).Announcing Kubetorch: Blazing Fast ML Development on KubernetesAnnouncing Kubetorch: Blazing Fast ML Development on Kubernetes
-
Vincent Moens reposted thisWe’ve always assumed stale and off-policy data hurts RL a lot — but our latest work shows the opposite. 🧠 M2PO (Second-Moment Trust Policy Optimization) reveals that even data stale by 256 model updates can train LLMs as effectively as on-policy RL, unlocking scalable and asynchronous RL scenarios. We also discovered a “Prosperity-before-Collapse” phenomenon: training without a trust region can temporarily outperform on-policy RL before divergence — suggesting stale data is surprisingly informative. M2PO tackles this by using a second-moment constraint and token-level masking, stabilizing off-policy learning while retaining high-entropy, high-value tokens that drive progress. 🚀 Stable. Efficient. Asynchronous. 🔗 Blog: https://lnkd.in/gQqCjG6Z 📄 Paper: https://lnkd.in/gf8sDkXc 💻 Code: https://lnkd.in/g4vTpFcZVincent Moens reposted this🤔𝐂𝐚𝐧 𝐰𝐞 𝐭𝐫𝐚𝐢𝐧 𝐑𝐋 𝐨𝐧 𝐋𝐋𝐌𝐬 𝐰𝐢𝐭𝐡 𝐞𝐱𝐭𝐫𝐞𝐦𝐞𝐥𝐲 𝐬𝐭𝐚𝐥𝐞 𝐝𝐚𝐭𝐚? 🚀Our latest study says 𝐘𝐄𝐒. 𝘚𝘵𝘢𝘭𝘦 𝘥𝘢𝘵𝘢 𝘤𝘢𝘯 𝘣𝘦 𝘢𝘴 𝘪𝘯𝘧𝘰𝘳𝘮𝘢𝘵𝘪𝘷𝘦 𝘢𝘴 𝘰𝘯-𝘱𝘰𝘭𝘪𝘤𝘺 𝘥𝘢𝘵𝘢, unlocking more scalable, efficient asynchronous RL for LLMs. We introduce M2PO, an off-policy RL algorithm that keeps training stable and performant even when using data 𝐬𝐭𝐚𝐥𝐞 𝐛𝐲 256 𝐦𝐨𝐝𝐞𝐥 𝐮𝐩𝐝𝐚𝐭𝐞𝐬. 🔗 Notion Blog: https://lnkd.in/gkx-k7xw 📄 Paper: https://lnkd.in/gK92KYVj 💻 GitHub: https://lnkd.in/gTpixWHu Joint work with Jiawei Zhao and Beidi Chen. #AI #LLM #LLMReasoning #GRPO #RL #reinforcementlearning
-
Vincent Moens shared thisI'll be talking at the Engineering nights tomorrow in London about the changing landscape in Reinforcement Learning Come if you want to chat! Register to the event here: https://luma.com/xhzzv62x Thanks Jules Belveze for the invitation!
-
Vincent Moens reposted thisVincent Moens reposted thisLightspeed is thrilled to have invested in Periodic Labs’s $300M seed round. This company is led by an incredible team of ex-OpenAI and DeepMind researchers, who are turning AI into an engine for scientific discovery. When I first met Co-Founders William (Liam) Fedus and Ekin Dogus Cubuk, I knew they were building something that could change many industries. Most AI companies focus on software. Periodic Labs is building the AI and the actual labs. Their system doesn't just theorize, it creates hypotheses, designs experiments, and then actually runs them through robotic systems. Every experiment, whether it succeeds or fails, teaches the AI something new. We're talking about accelerating scientific discovery in semiconductors, fusion energy, and sustainable materials, which are the building blocks for everything from better electronics to space travel. It's rare to find founders tackling problems this ambitious with this level of technical depth. Congrats to the entire team at Periodic Labs.
-
Vincent Moens liked thisVincent Moens liked thisI've been working on two new locomotion demos in AI4AnimationPy. One for humanoid locomotion where you can interactively choose one of many styles (e.g., "zombie", "neutral", "dinosaur") and one for quadruped locomotion. Besides PC you can also run them in your browser or on your phone in real-time, while using the same Python backend for all three. 🎮Try it out here: Web-Demos: https://lnkd.in/eXEwG3RG GitHub repository: https://lnkd.in/eGnGVarC Video: https://lnkd.in/eCDYx-4X #AI #Animation #MachineLearning #OpenSource
-
Vincent Moens liked thisVincent Moens liked thisExcited to share AIRA₂ — our next-generation AI Research Agents for ML that address key bottlenecks to scaling. Check out the paper: https://lnkd.in/exyCWtBE
-
Vincent Moens liked thisVincent Moens liked thisI'm excited to share that I joined Reflection AI last month on the Strategy and Operations team, where I'm working on Special Projects. In my first few weeks here, I've been energized by the mission and the people, and I'm already learning so much. We're hiring, so if you're interested in learning more, reach out!
-
Vincent Moens liked thisVincent Moens liked thisAre you heading to ICLR 2026 in Rio 🇧🇷? Meta will be hosting an evening to bring together folks across the AI community. It is a great chance to meet, exchange ideas, and expand your network. 📍 Rio de Janeiro 📅 April 24th, 2026 ⏰ 6:30–9:30 PM (BRT) Capacity is limited, so feel free to secure a spot here: 👉 https://lnkd.in/ejmsBRU7 #ICLR2026 #ICLR #LearningRepresentations #AI #ArtificialIntelligence #DeepLearning
-
Vincent Moens liked thisVincent Moens liked thisI am joining Resolve AI Labs as a founding member! 🎉 Resolve is building AI agents to solve some of the trickiest reasoning problems: debugging real-world production systems at scale for enterprise customers. Solving this requires multi-agent coordination, long horizons, lots of tool calls, and simulating entire fake companies to practice breaking them in fun ways. Real systems have petabytes of logs, thousands of services, and every “SEV0” is a unique and sparse event. Now that everyone is using coding agents, all that AI slop code has to run somewhere. Recently Dhruv Mahajan messaged me, “hey, I joined a startup”. He had just walked away from a Distinguished Research Scientist role at Meta, to join as Chief Scientist at Resolve. He was there many years and we overlapped on several projects including Llama. He convinced me to join: it’s a greenfield lab, extremely complex problems, growing customers, solid founders, no politics, and not immediately in the crosshairs of the frontier labs. At GrokStyle (acquired by Facebook) I held the SOTA for exact product recognition for several years. At Meta, I landed significant revenue working on research in both Commerce and Ads, then led the pre-training data team for Llama 3, 4, Emu, and what later became Muse. I enjoyed my time there, worked with some amazing people, and got promoted from Staff to Director over several years. At Resolve, I’ll be driving RL efforts to train new models in the “AI for prod” domain. I’ve been convinced that RL is the future for some time now - it has surpassed humans in some domains and the science is still super early. But I never had a chance to jump in until now. Dhruv and I sat down and wrote the vision of the Lab below. Take a look: https://lnkd.in/gAphN6_W (And yes, we’re hiring!)
-
Vincent Moens liked thisVincent Moens liked thisEvery day I receive invitations to "elite / VIP / invite-only / exclusive / c-level" events. Usually it is the same thing - a private room in a fancy restaurant, overpriced food, the same titles, the same small talk, and a lot of sales disguised as peer discussion. That's obviously a waste of time and this format is close to useless. The best conversations do not happen when 20 not hands-on people with similar titles are sitting around an expensive dinner table. The real value is when people with different backgrounds, from different domains, and with different responsibilities are in one room and can just talk to each other. On 21 May we are hosting the next London PyTorch Meetup at Revolut HQ. No fancy dinner, no face control, no sales. Just talks, finger food, drinks, and coffee from a standard office coffee machine. And most importantly, a chance for researchers, engineers, data scientists, founders, managers, and students to actually meet in one room. Happy to announce the first two speakers out of four: Anton Repushko, Head of AI/ML Research @ Revolut -> PRAGMA: Build Next-Generation Foundation Models in Finance Financial transactions share key properties with language: temporal ordering, contextual dependencies, and patterns emerging from surrounding context. This session presents a framework for building transaction foundation models using masked prediction and next-item forecasting to learn rich representations of customer behavior without labeled data. We'll cover our end-to-end approach—from tokenizing transactions into sequences to self-supervised objectives capturing spending rhythms, cross-merchant correlations, and anomaly patterns. Learn how this architecture drives improvements across fraud detection, credit risk, and behavior prediction while drastically reducing labeled data requirements. Nicolas Atienza, AI Scientist @ Neuralk -> Gradient Boosting to Tabular Foundation Models: What Changed, What Didn’t Tabular data is everywhere, yet progress has historically been incremental: better feature engineering, better boosting, better hyperparameter tuning. Tabular foundation models promise something different - learning a prior over datasets themselves, not just fitting parameters to one dataset at a time. This talk is aimed at data scientists and ML practitioners familiar with tabular modeling who want to understand whether these new models are hype or a genuine paradigm shift. PS: and 🤖 🤖 🤖 🤖 🤖 🤖 from Robomates [registration link is in the first comment]
-
Vincent Moens liked thisVincent Moens liked thisStill hiring for PhD candidates who are *specifically* excited about building and deploying RL systems for self-driving vehicles and other multi-agent planning settings. Shoot me an email if you think this is you and please help spread the word! Note this is an off-cycle hire.
-
Vincent Moens reacted on thisI am extremely excited about Markus Wulfmeier joining #nomagic as a Chief Scientist from Google DeepMind. AI is already changing the physical world and our production picking robots are one of the proofs of that. But there is much more of it yet to happen. Progress will come from research, data from production systems and the ability to apply the scientific method to achieve progress: iterating quickly while measuring results. Markus will enable us to accelerate this journey! Welcome on board! 🚀Vincent Moens reacted on thisWe are thrilled to announce that Markus Wulfmeier has joined Nomagic as our new Chief Scientist. Coming to us from Google DeepMind, Markus will spearhead our research into Visual-Language-Action (VLA) models. His leadership will be instrumental in accelerating the development of our Robotics Foundation Model (RFM), which represents the next generation of AI designed to help robots learn, adapt, and operate within real-world environments. This appointment marks a significant milestone in our commitment to creating physical AI solutions that go beyond simple perception to act with complete autonomy and reliability. We look forward to pushing the limits of embodied intelligence and transforming the future of warehouse automation. #PhysicalAI #Leadership #Nomagic #Robotics
-
Vincent Moens liked thisVincent Moens liked thisWeek 3 nearly in the books at Reflection. A few quick thoughts — and some roles we are hiring for. On the company: The mission — It's big, bold, and resonates with everyone. There's a clarity to it that has people genuinely aligned and moving in the same direction. That feels great. The vibes — Incredibly collaborative and open. There's a ton of work to do, and people are eager to pitch in across teams. Infra, product, research, policy, legal — it all just works seamlessly, with effectively no boundaries. The talent — Has eclipsed my wildest expectations. We're hiring some of the best people in the space from OpenAI, Anthropic, Google, and Meta, and that's attracting more top talent. It feels like things are snowballing in a way that's building a really strong team and culture. I can't talk yet about what we're working on, but I'll be doing talks and interviews soon — stay tuned. What we're hiring for: - Safety — a safety lead and a safety PM. Core to what we do, and to doing it openly. - Policy — so much happening at the global stage; hiring the best in regulatory, safety and preparedness, and more. - Legal — great work to be done on open source, commercial deals, and beyond. - Comms and community — events, hacks, technical blogs, publications. Looking for someone to help drive it. - SWEs/MLEs — we're building a ton in open source, and hiring engineers who are builders first: people who engage communities around their projects and evangelize in the ecosystem. Truly unique work, and unapologetically open source. - GPU infrastructure & compute — We're developing foundational models at very large scale, and hiring engineers to build GPU and compute infra across multiple clouds. - Post-training — Actively recruiting candidates to lead post-training. Given our roots, we're heavily invested in RL — so if large-scale RL, environments, and shaping foundational model behavior excites you, this could be a great fit. If any of this resonates, please submit at: https://lnkd.in/gjkZgE8f Cheers! Bonus: https://lnkd.in/gXgEyQp7
Experience & Education
-
Periodic Labs
****** ** ********* ***** *** ******** ****
-
****
******** ******** * ******* ********** * ******* ***********
-
****** ************ ******** * *********** **** ***
****** ******* ******** ******** *********
-
********** ********** ** *******
****** ** ********** * *** ************* *********** undefined
-
-
********** ********** ** *******
******** ****** ********
-
View Vincent’s full experience
See their title, tenure and more.
Already on LinkedIn? Sign in
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Publications
-
Learning and Forgetting Using Reinforced Bayesian Change Detection
Plos Computational Biology (Accepted)
-
The Hierarchical Adaptive Forgetting Variational Filter
2018 ICML Proceedings
Languages
-
English
Full professional proficiency
-
French
Native or bilingual proficiency
-
Italian
Native or bilingual proficiency
View Vincent’s full profile
-
See who you know in common
-
Get introduced
-
Contact Vincent directly
Other similar profiles
-
Sanj Ahilan
Sanj Ahilan
Building AI for Patents at Solve Intelligence. Trusted by over 400 IP teams, we assist legal professionals around the world in drafting, prosecution, litigation, invention harvesting and more.
3K followersUnited Kingdom -
Jifei Song
Jifei Song
Huawei Technologies Research & Development (UK) Ltd
1K followersLondon Area, United Kingdom
Explore more posts
-
mario carta
CNRS • 1K followers
Curious about how thermal inputs are encoded across species? Our new review article, “The neuronal circuits and cellular encoding of thermosensation", is now out in Nature Reviews Neuroscience: https://lnkd.in/eutSzx7Y Together with Mikkel Vestergaard and James Poulet, we discuss how rodents, primates, and insects process thermal stimuli, highlighting both shared mechanisms and key differences. Here you can find the full-text access to a view-only version: https://rdcu.be/eThU1
70
8 Comments -
Sheri Grach
ProLytics Consulting Group • 519 followers
I’m excited to share that our scoping review preprint has been published on medRxiv: A Scoping Review of Algorithmic Equity, Data Diversity, and Inclusive Design in the Transformer Era of Clinical NLP https://lnkd.in/eWvWM8h6 Many thanks to Abeer Badawi, Ph.D., Elham Dolatabadi, and Farah Ahmad, MBBS, MPH, PhD, for their collaboration, guidance, and support throughout this work. The rapid digitization of healthcare has positioned transformer-based NLP models as powerful tools for managing clinical text. However, their growing integration into practice raises important and unresolved questions about equity and inclusivity. This review synthesizes 56 studies (2017–2024) and examines how equity is addressed across three dimensions: ➡️ Algorithmic equity — fairness audits are largely post hoc and fragmented ➡️ Data diversity & representativeness — persistent underrepresentation creates what we define as Data Diversity Debt ➡️ Participatory design — observed in only 11% of studies, revealing a major gap beyond clinician involvement. To move beyond descriptive audits, we propose an equity-by-design roadmap to embed fairness, inclusivity, and accountability across the full lifecycle of clinical NLP systems—and to retire Data Diversity Debt. #HealthAI #ClinicalNLP #AlgorithmicEquity #DigitalHealthEquity #AIethics #ResponsibleAI #medRxiv
26
2 Comments -
Magda Dubois
AI Security Institute • 1K followers
It’s a big day for AI evaluations 🚀 The long awaited Inspect Scout has finally been released with all its great features. It's essentially a tool for in-depth analysis for AI transcripts. To make sure that people use it to its full potential, we aggregated best practices from multiple researchers (AISI, CAISI, METR, Princeton, RAND, Cambridge) and created a practical guide with concrete examples and code implementation 🔥 We tried to cover various applications (e.g., agentic evaluations, chatbot multi-turn conversations), different scoring methodologies, validation techniques etc. We really hope this will be useful to the community! Blog post: https://lnkd.in/eE3Tiq3A Best practices paper: https://lnkd.in/e_QxDDUu Inspect Scout: https://lnkd.in/etetUBdu
42
1 Comment -
Pasupathi Narayanan
Indian Institute of… • 1K followers
We are thrilled to introduce the "Open Orthophosphate Model," an AI and ML-driven framework set to transform the prediction of orthophosphate concentrations in UK water catchments. By utilising millions of historical observations and incorporating advanced molecular embeddings, correlations, and domain expertise, this model provides valuable insights into water quality. This innovative approach aims to aid the water quality monitoring, offering scalable, cost-effective solutions in addressing environmental, regulatory, and technical challenges. Please access the model under the GitHub location: https://lnkd.in/ek6a6QSn and please provide your comments. #WaterQuality #AI #MachineLearning #EnvironmentalTech #Innovation
17
-
Ben Batman
PBS • 810 followers
I've become interested in mechanistic interpretability recently and wanted to experiment with some of the latest innovations in the space. In a limited fashion, we can look inside a LLM and see what's going on. Using Sparse Autoencoders (SAEs) from Gemma Scope, this tool lets you peek under the hood of Google's Gemma 2 2B model and actually steer its behavior in real time. You can: - Browse and search 16K+ interpretable features by description - Visualize which features activate on any input text - Amplify or suppress specific features during generation and see how outputs change - Decompose predictions into per-feature logit contributions Built with TransformerLens, SAELens, and Gradio. Try it on Huggingface Spaces: https://lnkd.in/e88v8wym Code here: https://lnkd.in/e7dSS9WA #MechanisticInterpretability #AISafety #MachineLearning #NLP #OpenSource
6
-
Anna Seo Gyeong Choi
Rev • 484 followers
New paper accepted to EMNLP 2025 Findings! Not in my usual speech area but linguistic variation and fairness nonetheless 😃 Large language models (LLMs) are ubiquitous in modern day natural language processing. However, previous work has shown degraded LLM performance for under-represented English dialects. We analyze the effects of typifying “standard” American English language questions as non-“standard” dialectal variants on multiple choice question answering tasks and find up to a 20% reduction in accuracy. Additionally, we investigate the grammatical basis of under-performance in non-“standard” English questions. We find that individual grammatical rules have varied effects on performance, but some are more consequential than others: three specific grammar rules (existential “it”, zero copula, and y’all) can explain the majority of performance degradation observed in multiple dialects. We call for future work to investigate bias mitigation methods focused on individual, high-impact grammatical structures.
71
3 Comments -
Pragna Thotakura
Velsera • 1K followers
FunctionGemma, a new 270M parameter language model from Google, designed specifically for function calling and tool execution — not general chat. Why is FunctionGemma special? 🔹 Extremely lightweight FunctionGemma can run in ~550MB RAM, making it possible to deploy on laptops, edge devices, and even phones — all without large GPUs. 🔹 Purpose-built for function calling Unlike general LLMs, this model is trained for text-only tool calling, making it ideal for automation, agents, and API-driven workflows. 🔹 Easy local deployment The model is released in GGUF format, so it works seamlessly with llama.cpp and Ollama using quantized versions (Q4/Q5/Q8). 🔹 Fine-tuning ready from day one Thanks to Unsloth’s day-zero support, you can: Run FunctionGemma on CPU or GPU Fine-tune it locally using Colab notebooks Choose full fine-tuning or LoRA What can you fine-tune it for? FunctionGemma is meant to be customized for specific tasks, such as: Multi-turn function-calling agents Mobile actions (calendar events, reminders, flashlight, timers) Task-specific automation and workflows Lightweight, offline AI assistants This model clearly shows a shift toward small, focused, and efficient AI — where models are trained to do one thing well and can be adapted quickly to real-world use cases. Model (GGUF): https://lnkd.in/gxVuBhZA Docs & fine-tuning guides: https://lnkd.in/gchpAjUF Run and Deploy LLMs on your iOS or AndroidPhone: https://lnkd.in/grk72Bey
5
-
Ge Lei
635 followers
📢 Deep Matters: Foundations Speaker Announcement #DeepMatters ⭐️ Ronan Docherty, PhD student at Imperial College London 🎙 Talk: “Make Do and Mend: Leveraging Vision Transformers for Micrograph Segmentation” Ronan is a PhD student at Imperial College London, where he explores how modern machine learning can improve the segmentation of micrographs — a key step in materials analysis. Alongside his research, he works part-time at Polaron as a machine learning engineer, applying cutting-edge AI methods to real-world scientific problems. In his talk, Ronan will discuss how foundation models — particularly vision transformers trained on vast datasets of natural images — can be adapted for materials science. He’ll share strategies for fine-tuning these models to handle limited and out-of-distribution data, capturing intricate microstructural details essential for downstream materials characterization. By “making do and mending” with existing foundation models, Ronan demonstrates how researchers can achieve powerful results even without massive bespoke datasets — bridging the gap between modern computer vision and materials science. 📅 6 November 2025 📍 Google Cloud Startup Hub, London 🔗 https://lnkd.in/eVMsw_xk
6
-
Sagarnil Das
Meta • 8K followers
Meta is investing in the future of sound. 🔊 We’ve launched a new audio lab in the UK’s Ox-Cam corridor—one of Europe’s most dynamic research hubs—to push the boundaries of spatial audio, machine learning, and next-gen sound experiences. This lab will play a key role in building the future of human-computer interaction through intelligent audio. #MetaAI #AudioInnovation #ResearchAndDevelopment
12
-
Shaliza Panjwani
MedStar Washington Hospital… • 1K followers
Fascinating paper — a big step toward understanding self-supervised learning in a biologically realistic way. The idea of L5 activity reflecting prediction error without explicit labels is especially compelling. Raises important questions about how subcortical circuits might shape or refine these predictions.
-
Giancarlo Sperlì
Università degli Studi di… • 3K followers
If you are interested in a human-centered analysis investigating the acceptability of LLM tools in code translation, take a look at this paper! This work was done in collaboration with Anna Rita Fasolino Andrea Vignali Gabriele Dario De Siano. #LLMs #HumanCenteredAI #codetranslation
18
-
Teresa Zulueta-Coarasa
European Bioinformatics… • 656 followers
It was wonderful to chat with Oana Stroe about the recently published MIFA guidelines, which aim to standardise how we share AI-ready datasets for bioimage analysis. But even better than anything we could say ourselves are the testimonies from our users, who shared how the guidelines are already helping their research. Read the interview below!
13
-
Yoko Shimada
The Center for Global Digital… • 1K followers
If You Are “Building #AI” in #Health⚕️, This Is the Conversation We Need to Have 🧠 I have been having many conversations lately with organizations and #startup founders who say they are “working on AI.” Everyone wants to talk about AI. Everyone wants to do AI. What I often find missing, however, is a fundamental understanding that AI is nothing without #data - and more importantly, without #gooddata. 📊 As this post by Dr. Smisha Agarwal highlights so clearly, advanced #analytics cannot compensate for weak foundations. Working in the intersection of #health x #technology, I have seen this repeatedly. #Algorithms cannot fix #missingness, #bias, or poorly defined denominators. Just as importantly, we need to think more deeply about what datasets these algorithms were trained on in the first place. 🔍 🔍What blind spots are embedded in those data? 🔍How might those #blindspots shape or distort outputs? 🔍And why does that matter for real-world decisions in #healthcare? 📉 #Predictivemodels are also too often mistaken for #causal insight, and the consequence can be particularly serious in #publichealth and #healthcare contexts. ⚠️ And too rarely do we pause to ask whether simpler analytical approaches might answer the question just as well - or better. 📉➡️📈 This is especially concerning when we consider #gender #datagaps and other structural blind spots in health data. When #women, #marginalizedpopulations, or frontline realities and nuanced #contexts are poorly captured, AI will simply scale those blind spots with confidence. ⚖️ AI can be incredibly powerful - but only when paired with: 📈 strong data systems 🗄️ a deep understanding of the underlying data 📐solid #dataanalytics and #statistical foundations 🧠careful thinking about #causality versus #prediction and #humility about what the #evidence can (and cannot) tell us 🤍 As we move rapidly toward #AI-enabled #healthsolutions, the real work remains unglamorous but essential: building high-quality data, asking the right questions, and choosing methods that serve the problem - not the trend or the technology. 🌱 Thank you Dr. Smisha Agarwal for making these critical points visible and for grounding the AI conversation in #evidence, rigor, and humility. 🙏 #AI #DigitalHealth #DataScience #HealthSystems #PublicHealth #GenderData #EvidenceBased #HealthTech #populationhealth #machinelearning The Center for Global Digital Health Innovation Johns Hopkins Bloomberg School of Public Health #OODH
21
-
Meagan Lauber, PhD
SkyMedAI • 730 followers
I’m thrilled to share a new paper was recently published in the Alzheimer's Association®'s Journal, 𝘈𝘭𝘻𝘩𝘦𝘪𝘮𝘦𝘳'𝘴 𝘢𝘯𝘥 𝘋𝘦𝘮𝘦𝘯𝘵𝘪𝘢! 🧠 My previous PhD work used machine learning to explore the biology driving #Alzheimer's, this time I used #ML to validate digital tools aimed at making #dementia monitoring more accessible to patients! Collaborating with BU Alzheimer's Disease Research Center and Linus Health, we validated the Defense Automated Neurobehavioral Assessment (DANA) battery, a digital #cognitive test that can be administered to patients in their own homes, on their own smartphones! 𝘏𝘰𝘸 𝘥𝘪𝘨𝘪𝘵𝘢𝘭 𝘢𝘵 𝘩𝘰𝘮𝘦 𝘢𝘴𝘴𝘦𝘴𝘴𝘮𝘦𝘯𝘵𝘴 𝘳𝘦𝘷𝘰𝘭𝘶𝘵𝘪𝘰𝘯𝘪𝘻𝘦 𝘥𝘦𝘮𝘦𝘯𝘵𝘪𝘢 𝘤𝘢𝘳𝘦: 📲 𝐒𝐦𝐚𝐫𝐭𝐩𝐡𝐨𝐧𝐞 𝐓𝐞𝐬𝐭𝐢𝐧𝐠 = 𝐈𝐧𝐜𝐫𝐞𝐚𝐬𝐞𝐝 𝐀𝐜𝐜𝐞𝐬𝐬𝐢𝐛𝐢𝐥𝐢𝐭𝐲 ► Patients take tests at home on their own devices that they are familiar with. ► Removes travel and scheduling barriers, key obstacles for older adults with limited mobility. 📈 𝐓𝐫𝐚𝐜𝐤𝐢𝐧𝐠 𝐆𝐞𝐧𝐮𝐢𝐧𝐞 𝐂𝐨𝐠𝐧𝐢𝐭𝐢𝐯𝐞 𝐂𝐡𝐚𝐧𝐠𝐞𝐬 ► Repeat cognitive testing often causes “practice effects,” where people improve just by familiarity with test material. ► Our research validates that DANA is robust against these effects and can be used to capture true cognitive shifts over time. 🔎 𝐌𝐚𝐜𝐡𝐢𝐧𝐞 𝐋𝐞𝐚𝐫𝐧𝐢𝐧𝐠 𝐕𝐚𝐥𝐢𝐝𝐚𝐭𝐢𝐨𝐧 𝐟𝐨𝐫 𝐃𝐞𝐭𝐞𝐜𝐭𝐢𝐧𝐠 𝐂𝐨𝐠𝐧𝐢𝐭𝐢𝐯𝐞 𝐈𝐦𝐩𝐚𝐢𝐫𝐦𝐞𝐧𝐭 ► Random forest and logistic regression models trained on DANA responses achieved up to 71% accuracy at classifying cognitively impaired individuals. 🏥 𝐑𝐞𝐚𝐥 𝐖𝐨𝐫𝐥𝐝 𝐈𝐦𝐩𝐚𝐜𝐭 𝐨𝐧 𝐏𝐮𝐛𝐥𝐢𝐜 𝐇𝐞𝐚𝐥𝐭𝐡 ► Globally, there is a growing shortage of dementia specialists, coinciding with a rise in the number of patients living with dementia worldwide. ► Smart phone based digital assessment tools like DANA increase accessibility of dementia monitoring, no in-clinic visit needed! ► Remote, frequent cognitive assessments taken at home put control in the hands of patients and caregivers, enabling early detection and proactive management of brain health. It's a truly exciting time to be in the field of digital health! My PhD research has focused on leveraging cutting edge technology to both enhance our understanding of Alzheimer's disease biology and address real world health gaps in dementia care. I'm excited to keep integrating my expertise in health data science, neuroscience, and global health to improve health outcomes for neurodegenerative disease patients. 🌎 #digitalhealth #healthAI #research #machinelearning
72
6 Comments -
Sachin Awati
MPS Limited • 7K followers
🚨 Small number of poisoned samples can seriously compromise LLMs! I just read an eye-opening study by Anthropic, UK AISI, and The Alan Turing Institute titled “A Small Number of Samples Can Poison LLMs of Any Size.” Their core finding is astonishing: with as few as ~250 malicious documents, an attacker can successfully insert a “backdoor” into large language models (ranging from 600M to 13B parameters) — causing them to misbehave on specific trigger phrases — regardless of the total training data size or model scale. This marks a major shift from the earlier belief that an attacker must control a proportion of the dataset. Instead, this research shows that it’s the absolute number of poisoned samples that truly matters. The full technical paper — “Poisoning Attacks on LLMs Require a Near-constant Number of Poison Samples” — dives deeper into the experiments and implications. 📄 Read here: https://lnkd.in/gje8V-76 🎥 Watch a great explanation: https://lnkd.in/gMk68vj3 🙏 Thanks to Krish Naik for breaking it down so clearly!
34
2 Comments -
Mingxuan Zhang
Columbia University Irving… • 137 followers
Excited to share our new method SCONE (Scalable Contrastive Orientation for causal discovery), developed with my collaborators Khushi Desai and Sopho Kevlishvili , supervised by Elham Azizi! 🔸 SCONE tackles causal discovery under unknown soft interventions, a setting common in real-world systems such as gene regulatory responses to treatment 🔸 Introduces contrastive orientation rules across regimes, enabling recovery of causal edges that cannot be identified from any single regime alone 🔸 Scales causal discovery to large graphs by combining subset-level discovery with a global aggregation mechanism 🔸 Generalizes to out-of-distribution mechanisms, addressing a major limitation of many existing causal discovery methods 💡 Key idea: by comparing structural information across regimes, SCONE can identify edges that non-contrastive approaches fundamentally cannot recover, bridging the gap between theoretical guarantees and scalable practical methods. 📈 Experimentally, SCONE outperforms state-of-the-art methods across graph sizes and densities, including large graphs (100 nodes), even when other methods are given the true intervention targets—which SCONE does not require. 👉 Preprint: https://lnkd.in/e52n32vG 💻 Code: https://lnkd.in/eDpApMkY Huge thanks to my co-authors Khushi Desai, Sopho Kevlishvili and Elham Azizi for making this work possible! #AI #MachineLearning #CausalDiscovery #GraphLearning #Statistics #DataScience
52
2 Comments -
Edith B. Milanzi, PhD
Utrecht University • 3K followers
2025 was quite the year, but thanks to some down time, I’ve been able to catch up on recent work on AI and data ethics in the context of LMIC systems. I found the RAi UK white paper on Responsible AI in LMICs interesting for how it explores AI adoption in contexts where data systems are still immature and local priorities are overridden by global frameworks. It outlines the reality many of us see in health and social systems on the ground and raises questions about who innovation is really being designed for. UNESCO also updated the Playbook on tackling gender bias and harms in AI. I liked how practical this work is. It shows how the same inputs can still produce biased outputs. In settings where women’s data is poorly captured, these biases amplify inequalities and actively make them worse. Reading these back to back brought into focus the same clear message. Ethical AI is about about better models yes, but more so about data, the systems that produce it, and the decisions that shape its use. Adding AI does not improve decision making if data is incomplete and governance is patchy. It is just the same problems just dressed up as innovation. Curious about what 2026 will bring as this space keeps changing. 🔗 RAi UK White Paper https://lnkd.in/eiKNStaJ 🔗 UNESCO Red Teaming Playbook https://lnkd.in/ekW-M6u7
22
1 Comment -
Ganesh Venkatesh
Waymo • 2K followers
I'm incredibly proud to announce a productive NeurIPS for our Post-training AI Research team. We had two papers on test-time scaling accepted into the Workshop on Efficient Reasoning, contributing to a fantastic ~6 acceptances for the wider Applied AI Research @ Cerebras. Both of our papers are now available on ArXiv: - The Conductor and the Engine: A Path Towards Co-Designed Reasoning. Link - https://lnkd.in/g5Dnnz6C. - Calibrated Reasoning: An Explanatory Verifier for Dynamic and Efficient Problem-Solving. Link - https://lnkd.in/gH2jTGeP. In the spirit of collaboration, we are also open-sourcing our updated CePO flow as part of our OptiLLM library on GitHub. This is the same methodology that achieved top scores on the Artificial Analysis Leaderboard. Link - https://lnkd.in/gd6anwFN Congratulations to the team on this incredible milestone: Anisha Garg, David Bick, Engin Tekin, Michael Wang, Pawel Filipczuk, Amaan Dhada, Yash More, Nishit N. and rest of the Post-training Team! Looking Ahead A key enabler for these results and our upcoming work in Coding Agents has been using Reinforcement Learning (RL) to provide LLMs with new, specialized expertise, which makes them highly amenable to test-time compute scaling. This brings me to our next exciting step... 🚀 Announcing Limited Early Access to our RL Service! 🚀 Is your team excited by the potential of powerful LLM models — closed source like GPT-5/Claude/Gemini or Open-source like Qwen3, GPT-OSS — but frustrated when they fail that "last mile" on your specific, critical tasks? We are opening up a limited early access program to help you solve this. Our RL service is designed to transform general models into world-class experts for your unique domain. If you're interested in building an AI system that is an expert at solving your tasks, reach out to me to see if you're a fit for the program. 📧 Email: ganesh.venkatesh@cerebras.net When you reach out, please include: - Subject: "RL Service Early Access" - Body: A brief description of your application and the "last-mile" challenges you're facing. Looking forward to post-training your problems away!
122
1 Comment
Explore collaborative articles
We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
Explore More