Sign in to view Jiarui’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Jiarui’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
London, England, United Kingdom
Sign in to view Jiarui’s full profile
Jiarui can introduce you to 4 people at PyData London
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
2K followers
500+ connections
Sign in to view Jiarui’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Jiarui
Jiarui can introduce you to 4 people at PyData London
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Jiarui
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Sign in to view Jiarui’s full profile
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
About
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
Activity
2K followers
-
Jiarui Cao, PhD shared thisI was lucky to stay a few nights when I was doing my Math PhD in Warwick. It was such an experience to sleep in a bedroom full of blackboards! Sadly, no mathematical inspiration came to me in my sleep and the blackboards were not fully used. 😂Jiarui Cao, PhD shared thisA short walk from our campus Piazza are some of Warwick's oldest architectural icons - the Maths Houses 🏠 Built in 1969 to solidify Warwick as a global hub of mathematical research by accommodating visiting mathematics, the buildings each famously have a study space with a continuous blackboard nearby so that mathematicians could work undisturbed. However this design also serves family life, too! As Warwick Mathematical Institute's Founder, Sir Christopher Zeeman, had a very clear brief for the architects bringing the project to life: "Put the blackboards low enough for small children to use the bottom bit.” Today the Maths Houses are Grade II* listed and considered as ‘buildings of particular national importance or special interest’.
-
Jiarui Cao, PhD reposted thisJiarui Cao, PhD reposted thisMeetup #78: PyData London 106th Meetup Tuesday, 7th April 2026 (#proofofnetwork) Nicoleta Lazar explored the idea of query federation which is all about being able to analyse data from multiple sources at once without having to copy it into one place first, making data access faster, simpler, and more democratic for everyday users. Instead of analysts manually stitching data together in warehouses, a smart query engine does the heavy lifting, although this comes with trade-offs like slower performance if one data source lags, plus challenges around security, network limits, and keeping everything up to date. She explained that there are two main approaches: either a flexible, all-in-one query layer or a more controlled system built into databases like StarRocks and highlighted how modern setups using cloud storage and formats like Apache Iceberg are shaping the future, where the goal is to balance speed, cost, and reliability while making complex data feel easy to use. Ben Guerin shared the story behind his personal project, ismypubfucked.com, to highlight the growing pressures facing British pubs. It starts with a simple idea: pubs might be getting short-term relief from government business rates, but the bigger picture is still worrying, with rising costs and that support set to expire in 2029, so the goal became helping people find and support the pubs that need it most. To make this happen, Ben began by sketching out the idea in plain language, gradually refining it into a clear plan with the help of AI, before turning it into a working project. He then used Python to pull in massive datasets from official valuation lists, clean up messy and confusing raw data, and compare how pub values have changed over time, effectively revealing which ones are under the most strain. From there, he mapped every pub across Great Britain, solving tricky issues like mismatched names and missing locations, until he had a fully verified geographical dataset. The result is something he calls the “Fucked Pub Index”, a simple score that shows how much financial pressure each pub is under, ranging from doing fine to seriously struggling. Finally, he loaded all of this into a database in a way that’s fast and seamless, ending up with over 46,000 pubs analysed and ready to explore, turning complex data into a clear, relatable way for people to understand and support their local pubs. Rhys Green introduced ChatPT, an AI-powered fitness app that started as a personal project but is aiming to become an affordable, always-available alternative to traditional online coaching. It works like a conversation, guiding you through creating personalised training and nutrition plans, while also letting you track progress, revisit plans, and explore exercises with helpful instructions and videos. The idea is to make fitness planning feel simple and natural, bringing everything, from coaching to daily logging into one easy-to-use platform.
-
Jiarui Cao, PhD reposted thisWhat an incredible line up of keynotes! Make sure to check out the schedule and get your tickets today!! https://lnkd.in/eCCSUn8UJiarui Cao, PhD reposted this🚨 PyData London 2026 Keynote Speakers Announced 🚨 We’re thrilled to unveil an incredible lineup of keynote speakers for PyData London 2026, bringing together some of the most influential voices in data, AI, and engineering: Sam Colvin – Founder of PyDantic Inc Rachel Lee-Nabors – Agentic Web Leader Jeremiah Lowin – Founder & CEO of Prefect Martin O'Reilly – Director of Research Engineering at the Alan Turing Institute New for 2026: We’re kicking things off early with a special keynote from Sam Colvin on Tutorial Day, Friday, June 5, 2026 — setting the tone for an unforgettable weekend of learning and connection. Purchase your 3-day pass today — you don’t want to miss out! https://hubs.la/Q048kj170
-
Jiarui Cao, PhD reposted thisEveryone's talking about AI. Not everyone understands it. My secret to staying ahead? Surrounding myself with the people who actually build it. 🐍 PyData London is three days of real talk — no hype, no buzzwords, just data scientists, engineers and developers sharing what actually works. 📅 June 5–7, 2026 📍 Convene Sancroft, St. Paul's, London 🎟️ Ready to stop feeling behind? 👉 https://lnkd.in/esVfcfka See you there! #PyData #PyDataLondon #DataScience #AI #Python
-
Jiarui Cao, PhD shared thisLondon has one of the most vibrant data & AI communities in the world, and PyData has long been part of that ecosystem. I’m excited to be helping organise the PyData London Conference 2026, part of the global PyData / NumFOCUS open-source community supporting projects such as NumPy, Pandas, Scipy, and jupyter. 📅 5–7 June 2026 Each year the conference brings together a highly technical audience of data scientists, ML engineers, researchers, and technical leaders from startups, big tech, finance, and academia. Previous PyData London conferences have been supported by companies such as NVIDIA, Bloomberg, Databricks, Man Group and Anaconda, Inc. We are currently speaking with organisations interested in sponsoring the 2026 conference. For companies building AI products, data platforms, ML infrastructure, or developer tools, PyData London is a great way to engage with the community and connect with experienced practitioners. If your company might be interested, feel free to message me directly. And if you introduce a company that ultimately becomes a sponsor, we’d be happy to offer a complimentary conference ticket as a small thank-you. Looking forward to another great PyData London this year.
-
Jiarui Cao, PhD reposted thisJiarui Cao, PhD reposted thisMeetup #76: PyData London 105th meetup Tuesday, 3rd March 2026 (#proofofnetwork) Yusuf Ganiyu from AstraZeneca explored how data engineering is moving from AI as a simple tool to AI as a true teammate. Instead of the usual chatbot flow: ask a question, get a suggestion, copy and paste the code to run it, agentic workflows start with a goal, reason through the problem, take action, review the results, and keep iterating until it’s fixed. The agent consists of a reasoning engine to think, a planner to split tasks up, tools to take action like running queries or checking logs, and memory to keep track of context. These agents can monitor data quality more intelligently, retry minor failures automatically, suggest fixes for bigger issues, and escalate serious problems with full diagnostics attached. Yusuf even showed how an agent can debug pipelines, detect schema changes, and open pull requests, but with guardrails in place. That said, AI still can’t truly understand business context, company politics, or complex trade-offs, which is why maturity matters. So start small, give read-only access to agents first, and remember that your human judgement is still your biggest advantage. Jethro Reeve showcase his brilliant personal project called “Explore the Kingdom” (https://lnkd.in/e-xr9_dU), built using Cursor, which lets you explore the UK’s social and political landscape in a way that feels almost interactive and alive. You can compare house prices, rent, and energy costs across all 650 constituencies to see where you might fit in, dig into the 2024 election results and historical swings, and visualise income, deprivation, and education data. Behind the scenes, he pulled together 15 government data sources and mapped everything to constituency level using clever aggregation methods, fallback strategies, and clear provenance tracking to keep the data trustworthy. He also created synthetic profiles and fast heat maps to make the experience smooth, plus a nowcasting pipeline that pulls from RSS feeds and research to simulate how events might shift trends. Built with a modern web stack and deployed online, the whole system shows how today’s AI coding tools can act almost like a data engineer; browsing data, querying databases, and following defined workflows without opening a single spreadsheet. Latest AI models have crossed an important threshold in capability, but agentic tools are still rough around the edges, so you need strong guardrails and a healthy dose of caution. Stelios Christodoulou showed how to use GitHub Codespaces as a cloud-based development environment to perform data analysis without setting anything up on your own machine. By adding simple configuration files to your project you can create a consistent, ready-to-go workspace for everyone on the team. Comparing with other tools like Google Colab and Kaggle, there is a free option for individuals (up to 60 hours/mo) before moving to straightforward pay-as-you-go pricing.
-
Jiarui Cao, PhD reposted thisJiarui Cao, PhD reposted this[DEADLINE EXTENSION] The PyData London Diversity Scholarship application deadline has been extended to March 9! So, if you are interested in applying, you have a few extra days to get it in. Learn more at https://hubs.la/Q045dd400
-
Jiarui Cao, PhD shared thisJoin me at PyData London - 105rd Meetup https://lnkd.in/erBFyypUPyData London - 105rd Meetup, Tue, Mar 3, 2026, 6:30 PM | MeetupPyData London - 105rd Meetup, Tue, Mar 3, 2026, 6:30 PM | Meetup
-
Jiarui Cao, PhD reposted thisJiarui Cao, PhD reposted thisMeetup #74: PyData London 104th Meetup Tuesday, 3rd Feb 2026 (#proofofnetwork) David Salvador-Jasin, PhD introduced SVD-ROM, an open-source Python package that makes it possible to simplify and analyse massive datasets without needing supercomputers. The idea is surprisingly intuitive: many complex systems are driven by just a few dominant patterns, and by using maths techniques like Singular Value Decomposition, SVD-ROM compresses huge data down to these key signals while keeping the important behaviour. Built on tools like Dask and Xarray, it can run on a laptop, a cluster, or the cloud, making large-scale forecasting faster, cheaper, and more explainable; a big win for anyone working with modern, data-heavy systems. Anton Nazaruk walked through how to go from zero to large-scale AI training on AWS without drowning in infrastructure complexity. He explained why GPUs are hard and expensive to manage. Capacity limits, failures, and runaway costs. He also showed how AWS options range from DIY EC2 setups to managed SageMaker tools like HyperPod. HyperPod acts like a resilient training cluster that automatically replaces failed nodes and resumes jobs, making it ideal for long, production-grade training runs. AWS can handle the hard infrastructure bits, but you still own your code, data, and metrics. With a clear blueprint for compute, networking, storage, and observability, large-scale training becomes repeatable rather than painful. Mojtaba Kargar shared how his team built an AI-powered platform for personalised health reports, helping longevity clinics shift from reactive care to proactive medicine. With over 10,000 data points per patient, the real challenge was managing complexity, which they solved using specialised AI agents that validate each other and share context rather than relying on one central brain. Key learnings: decentralised design, human oversight, and continuous validation work, while monolithic prompts and central control don’t. Hugh Evans shared a personal project where he mapped the global PyData community using web scraping and data wrangling. He built a PyData meetup mapper that makes it easy to see how many meetups exist in each city and which ones are active or inactive. With simple colour-coded maps, it turns a messy global picture into something clear, visual, and genuinely useful.
-
Jiarui Cao, PhD liked thisJiarui Cao, PhD liked thisMeetup #79: AI Signals 30th Meetup Fri, 17th April 2026 (#proofofnetwork) ✨ Anni Chen from Twilio explored the exciting shift toward using voice as a primary interface, highlighting how speaking and listening are simply much faster and more natural for us than typing. She walked us through some behind-the-scenes magic required to build a real-time AI voice agent, explaining that while the architecture involves several moving parts, like detecting when a human interrupts or orchestrating the flow between the AI's brain and its voice, the biggest hurdles are often lag and timing. To solve this, Twilio introduced ConversationRelay, a clever tool that lets developers "bring their own AI" (BYOA) to a phone call while Twilio handles the heavy lifting of translating speech to text (and vice versa) across thousands of different voices. If you wanted to build one yourself, you’d essentially just need a Twilio account, your LLM of choice, and a bit of personality enhancement to make the agent sound human (i.e. specifying tones, styles, cadence etc). Anni wrapped things up with a live demo where the audience could actually phone the AI to give feedback, showing that with the right streaming and error-handling practices, voice agents can handle everything from language switching to complex customer queries. Peter Houghton started with showing how test automation and software quality are different, and hence need separate solvable approaches. In rigid but structured environments like banking, where systems are governed by a number of complex rules and slow waterfall-esque development cycles, we can use AI agents to speed things up by hunting for the opposites. By feeding an AI the official rules and asking it to intentionally create data that breaks them, we can discover hidden errors in the code or even the rules themselves much faster than a human could. The core workflow involves turning those rules into a validator, using AI to generate messages that test every edge case, and then building apps to process them, ensuring everything is grounded in reality through spot checks and cross-referencing. Ultimately, his message was about building an investigative capability rather than just a simple test suite. By using reasoning models to work backwards from end results to initial inputs, teams can bypass traditional bottlenecks and build much more reliable systems with AI as a partner in quality control. Anand Rawat presented the concept of Autonomous DataOps, which is essentially like giving a data pipeline its own brain so it can monitor itself and fix problems without waiting for a human to step in. Most AI systems struggle when data changes unexpectedly or signals go missing, but by using specialised agents that can observe, reason, and act, these systems can actually become self-healing. Whether it’s managing credit risks or monitoring health systems, the goal here is to build smarter infrastructure that catches errors in real time.
-
Jiarui Cao, PhD liked thisThis is your very last chance to submit a talk to PyData Amsterdam 2026 (aka one of the coolest open-source conference in Europe, and I'm not biased at all! 😉) If you have a great story to share, we would love to hear about it 🫶Jiarui Cao, PhD liked this🐣 With Easter weekend just behind us, we want to make sure everyone had a fair chance to submit their talk proposal! We're extending our Call for Proposals deadline to the end of the week (April 19th 23:59 CEST) so if the holiday got in the way, you've still got time. PyData Amsterdam is built by this community and your voice matters, so we want to hear from you! 🗓️ New deadline: April 19th (23:59 CEST) 📍 PyData Amsterdam 👇 Submit here: https://lnkd.in/eBEWBukP #PyData #PDAmsterdam2026 #CFP
-
Jiarui Cao, PhD liked this👉 Exciting opportunity—join us and be part of our growth journey!Jiarui Cao, PhD liked thisLooking for a strong Director to help lead our next phase of growth
-
Jiarui Cao, PhD liked thisJiarui Cao, PhD liked this😍 𝗠𝗲𝗲𝘁 𝗼𝘂𝗿 𝗸𝗲𝘆𝗻𝗼𝘁𝗲 𝘀𝗽𝗲𝗮𝗸𝗲𝗿 - Ines Montani 𝘊𝘰-𝘧𝘰𝘶𝘯𝘥𝘦𝘳 & 𝘊𝘌𝘖 𝘰𝘧 Explosion. Core developer of spaCy - one of the most widely used open-source NLP libraries in Python. Working at the intersection of AI, Natural Language Processing, and real-world applications, she has been shaping how developers build and use language technologies today. 🔸 Her work focuses on practical, production-ready NLP - from tooling to systems that actually work beyond prototypes. This is the level we’re bringing to Yerevan. _ 𝗣𝘆𝗗𝗮𝘁𝗮 & 𝗣𝘆𝗖𝗼𝗻 𝗬𝗲𝗿𝗲𝘃𝗮𝗻 𝟮𝟬𝟮𝟲 𝗝𝘂𝗹𝘆 𝟮𝟰–𝟮𝟱 Call for Proposals is open. 👉 Deadline: April 19. Apply to become a speaker: www.pycon.am PyData Yerevan, PyData Global, PYerevan, PyCon, VXSoft LLC, AUA Akian College of Science and Engineering, American University of Armenia, Plat.AI, Ines Montani, Habet Madoyan, Karen Javadyan, Ruben Manukyan, Suren Poghosyan, Mane Vardanyan #pydatapycon2026 #pydatayerevan #pyerevan #pycon2026 #python #ML #AI #computervision #techevents #conference #AUA #ACSE #spaCy
-
Jiarui Cao, PhD liked thisJiarui Cao, PhD liked thisWe're excited to be at PyCon DE & PyData Darmstadt 2026! Be sure to swing by the NumFOCUS booth to say hi and talk with us about helping support the open source scientific computing community. 🧡
-
Jiarui Cao, PhD liked thisJiarui Cao, PhD liked thisI am urgently hiring a Senior Individual Contributor (Data Science / LLM) in Bucharest, with strong hands-on LLM experience and solid DS/ML background. High-impact role with real ownership, working on production AI use cases that drive business outcomes. The interview process is fast and efficient (1-2 weeks). For highly experienced candidates, there is potential flexibility to hire at a higher band. If you have connections in Bucharest, I would greatly appreciate your help in sharing this opportunity within your network. 🙏
-
Jiarui Cao, PhD liked thisJiarui Cao, PhD liked thisWe're proud to share that we have a record number of subjects in the top 50 and top 100 of the QS World University Rankings by Subject 🎉 We now proudly offer: • 10 subjects ranked in the global top 50 • 20 subjects in the global top 100 Every part of our teaching community - the Faculty of Arts; Faculty of Science, Engineering and Medicine; and Faculty of Social Sciences - are represented in the top 50 subjects. This places us among an elite number of global universities demonstrating both breadth and depth of excellence across disciplines. “This recognition reaffirms the strength of our academic community, which is outward‑looking and committed to shaping solutions that have a real impact. Warwick has always been a university that leads with confidence and clarity. These results demonstrate once again our determination to push boundaries, elevate global knowledge and point the way ahead so that, together, we can make a better world.” - Professor Michael Scott, Pro-Vice-Chancellor (International) and Institutional Lead for Rankings. Find out more: https://bit.ly/4t5uBx1
Experience & Education
-
Barclays
**** ***** ********* * ** ********* **** *********
-
****** ******
** *********
-
******** *******
****** **** *********
-
********** ** *******
****** ** ********** ***** ******* *********** undefined
-
-
************ **********
****** ** ******* ******* ************ ******* *********** **** * ** ******
-
View Jiarui’s full experience
See their title, tenure and more.
Welcome back
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
New to LinkedIn? Join now
or
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Licenses & Certifications
Volunteer Experience
-
Committee Member
PyData London
- Present 1 year 11 months
Science and Technology
PyData London is a community backed by NUMFOCUS and ran by volunteers. We have Meetup in the evening on every month’s first Tuesday. Check and join us!
Honors & Awards
-
Winning team of Barclays' GenAI Hackathon
Barclays
Member of a team selected as one of the top 5 teams out of 280+ participating teams globally. Presented to Group CEO and Group CIO. Received funding to transfer Hackathon idea into project in production.
-
National Excellent Student Study Abroad Scholarship
China Scholarship Council
- National level award for PhD student.
- Only 35 awarded in UK (less than 1%). Only 500 awarded globally. -
Postgraduate Mathematics Project Prize
Loughborough University
Awarded to the best Master research project in School of Mathematics (only one prize awarded annually).
-
Chancellor's International Scholarship
The University of Warwick
Covers full fees (oversea student level) and stipend for 3.5 years PhD study.
The most competitive scholarship in University of Warwick. Only 25 are awarded annually from all overseas applicants across all disciplines.
-
Outstanding Exchange Student Scholarship
Loughborough University
Awarded to outstanding exchange students from China (less than 5% awarded)
Organizations
-
PyData London
Committee member
- PresentOne of London chapter organizers, where I help to facilitate monthly meet-ups and annual conferences.
-
Association for Computing Machinery
Member
- Present
View Jiarui’s full profile
-
See who you know in common
-
Get introduced
-
Contact Jiarui directly
Other similar profiles
Explore more posts
-
Dimitrios Nikolaidis
Bank Information Systems SA • 823 followers
𝗘𝘅𝗽𝗹𝗮𝗶𝗻𝗮𝗯𝗶𝗹𝗶𝘁𝘆 𝘄𝗼𝗿𝗸𝘀 — 𝘄𝗶𝘁𝗵𝗶𝗻 𝗹𝗶𝗺𝗶𝘁𝘀 Over the past decade, post-hoc explainability has become a widely adopted practice: Feature importance scores SHAP values Local surrogate models They provide insight into how predictions are generated. Yet they can: • Shift under small model perturbations • Appear stable even when models change materially • Be manipulated without altering predictive performance • Capture statistical associations rather than causal structure This does not invalidate interpretability tools. It clarifies what they are and what they are not. Most explanation tooling operates downstream of inference. Governance challenges emerge upstream, in how systems are deployed, updated, and assigned clear ownership If accountability depends primarily on dashboards, it becomes difficult to sustain over time. Part 3 will outline the architectural design choices that make AI systems defensible, not just explainable.
15
-
Caleb Severn
Capco • 2K followers
This week: https://lnkd.in/ejz9xxxp (Apple & UCSB) DEEPAMBIGQA: Ambiguous Multi-hop Questions for Benchmarking LLM Answer Completeness This paper looks at how LLMs can answer questions better by combining them with knowledge graphs. Firstly they generate question–answer pairs using multi-hop data from Wikidata, requiring more than just simple lookups or semantic similarity, for example the question “Which actors in the film Heat have won an Oscar?”. They then test three versions of LLMs answering questions: firstly a traditional LLM trying to answer directly, then one using query expansion, where the model reformulates the question using the entities within the graph to help disambiguate information. This means it can ignore songs or albums and focus on films, expanding the question to give more context → “Which actors in the 1995 film Heat directed by Michael Mann have won an Oscar?”, and finally an LLM that tries to write and execute SPARQL queries on the graph. The outcome? Disambiguating the entities using graph classes was the most helpful, increasing recall by about 13 percentage points, but the LLM-written SPARQL queries failed in more than 40% of cases. How can this be improved? Wikidata itself doesn’t explicitly define domains and ranges for most relations, so property use is ambiguous. The authors address this by deriving a simplified schema (an inferred ontology) from instance data, which helps constrain reasoning paths. Having more explicit domains and ranges, or using subproperties or OWL restrictions to describe where relationships can occur, would likely help LLMs generate more valid queries. Another benefit of a clear ontology is that it enables schema-based validation: you can check whether the generated SPARQL fits the expected pattern. For example, “You used a relationship that expects X as a subject but you used Z instead.” That feedback loop could help the LLM select the correct property next time. It’s great to see benchmarking that pushes beyond simple fact lookup and highlights the value of classifying entities and applying ontological modelling alongside LLMs. Great work Jiabao Ji, Min Li, Priyanshu Kumar, Shiyu Chang, Saloni Potdar #KnowledgeGraph #Apple #LLM #GenerativeAI #RDF #Wikidata
2
1 Comment -
Jon Minton
Smith+Nephew • 961 followers
Another short blog post, and my first hand-drawn post for a while. https://lnkd.in/eYDjbw7K This post attempts to explain differences between the chatbot-style LLMs most people are familiar with, and Agentic AI. In brief: With chatbots Human and Bot interface through a chat window. With Agentic AI we need to think about Human, Bot and Machine as distinct elements, with Human and Bot both sharing access to (and so indirectly communicating via) the Machine, as well as the chat window. This one simple difference is key to why AAI can be so much more powerful than AI through chat window alone.
4
-
Oscar Barlow
2K followers
Get small 🤏 "Our key empirical finding: Determinism is not universal across model architectures. Evaluating five models across 480 runs (n=16 per condition), we demonstrate that well-engineered smaller mod- els (7-8B parameters) achieve perfect output consistency, while 120B parameter models fail at 12.5% consistency even with identi- cal configuration (𝑇=0.0, greedy decoding, fixed seeds). This inverse correlation between model scale and determinism fundamentally alters deployment strategies for regulated applications, where audit requirements mandate reproducibility over raw capability." https://lnkd.in/emzbF4RC
11
1 Comment -
Abhishek Sehgal
Medtwin • 5K followers
We just released our new arXiv paper: “Structure and Diversity Aware Context Bubble Construction for Enterprise Retrieval Augmented Systems”. arXiv: https://lnkd.in/gJY9USty Many RAG pipelines still default to Top K. In real enterprise docs (multi sheet Excel, scopes, contracts, technical packs) that causes fragmentation, duplicate evidence, and wasted tokens. Context Bubbles treat context assembly as an explicit constrained selection problem. -> Structure aware scoring -> Diversity and redundancy control -> Strict token and per section budgets -> A full retrieval trace for auditability and deterministic tuning In our case study we cut context from 780 tokens to 214 while improving multi section coverage and reducing overlap. Guess what? There is allot more we did behind the scenes and this paper is the first in a series on building more ‘aware’ enterprise retrieval systems, where we combine structure, diversity constraints, and auditable selection with a broader memory and agent based architecture we are rolling out. The coolest thing we have built is our AI Engine which has created it's own code base and language and is taking the route of quantum inspired structures. This is already running in our ERP workflows and our medical research pipelines to reduce context bloat and improve coverage and citation faithfulness under limited windows. Keen to hear thoughts from anyone building RAG on structured enterprise data. #RAG #LLM #EnterpriseAI #InformationRetrieval #AIEngineering
5
-
Ralph Lozano
JPMorganChase • 431 followers
DeepSeek’s recent paper introduces "Hyper-Connections" (HC) as a method to enhance model capacity by expanding the residual stream into multiple parallel paths. This approach increases information capacity; however, the unrestricted mixing of these streams disrupts the identity mapping property, resulting in signal instability during training. I have written a blog post that analyzes their solution, mHC.
18
1 Comment -
Srinivasa Rao Aravilli
Capital One • 3K followers
We published a survey kind of paper on "From Attention to Disaggregation: Tracing the Evolution of LLM Inference". It will be invaluable for anyone interested in learning more about LLM Inference, its comparison with the CAP theorem (a concept familiar to those who have worked with distributed systems), and the state-of-the-art research in Generative AI inference https://lnkd.in/gjVe5P8J
77
-
Jonathan Lellouche
Murex • 2K followers
🧠 𝗠𝗟 𝗳𝗼𝗿 𝗣𝗗𝗘𝘀 = 𝗘𝗻𝗱 𝗼𝗳 𝘁𝗵𝗲 𝗖𝘂𝗿𝘀𝗲 𝗼𝗳 𝗗𝗶𝗺𝗲𝗻𝘀𝗶𝗼𝗻𝗮𝗹𝗶𝘁𝘆 📊🔥 High-dimensional PDEs, long considered intractable, are now being tackled head-on by Deep Learning. Once seen as a numerical bottleneck, PDE resolution via ML enables mesh-free, scalable and generalizable solvers – from option pricing to physics. — 🧠 𝗗𝗲𝗲𝗽 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴 𝗳𝗼𝗿 𝗣𝗗𝗘𝘀 𝗶𝗻 𝗢𝗽𝘁𝗶𝗼𝗻 𝗣𝗿𝗶𝗰𝗶𝗻𝗴 – 𝗧𝗶𝗺𝗲 𝘁𝗼 𝗧𝗮𝗸𝗲 𝗘𝗿𝗿𝗼𝗿 𝗔𝗻𝗮𝗹𝘆𝘀𝗶𝘀 𝗦𝗲𝗿𝗶𝗼𝘂𝘀𝗹𝘆 📉📈🧮 𝗡𝗲𝘄 𝗽𝗮𝗽𝗲𝗿 𝗯𝘆 𝗝. 𝗥𝗼𝘂 (𝗠𝗮𝘆 𝟮𝟬𝟮𝟱) proposes a systematic empirical error analysis of two deep PDE solvers for option pricing: → 𝐃𝐞𝐞𝐩 𝐆𝐚𝐥𝐞𝐫𝐤𝐢𝐧 𝐌𝐞𝐭𝐡𝐨𝐝 (𝐃𝐆𝐌) → 𝐓𝐢𝐦𝐞 𝐃𝐞𝐞𝐩 𝐆𝐫𝐚𝐝𝐢𝐞𝐧𝐭 𝐅𝐥𝐨𝐰 (𝐓𝐃𝐆𝐅) ✔ Evaluated on both Black–Scholes and Heston models ✔ Convergence rates for 5 key training parameters (layers, nodes, samples, time steps…) ✔ Surprising insight: number of samples has almost no effect on DGM accuracy ✔ Clear edge of TDGF with second-order discretization — 🔍 𝗪𝗵𝗲𝗿𝗲 𝗱𝗼𝗲𝘀 𝗶𝘁 𝗳𝗶𝘁 𝗶𝗻 𝘁𝗵𝗲 𝗱𝗲𝗲𝗽 𝗣𝗗𝗘 𝗹𝗮𝗻𝗱𝘀𝗰𝗮𝗽𝗲? ➡ 𝐏𝐈𝐍𝐍𝐬 (Physics-Informed Neural Networks) Proposed by Raissi et al. (2019), PINNs solve PDEs by penalizing residuals directly in the loss function. Ideal for high-dimensional problems – but can be unstable or slow to converge. Use case: widely supported in 𝗡𝗩𝗜𝗗𝗜𝗔 𝗠𝗼𝗱𝘂𝗹𝘂𝘀, a powerful PINN framework for scientific ML. ➡ 𝐃𝐞𝐞𝐩 𝐁𝐒𝐃𝐄 𝐌𝐞𝐭𝐡𝐨𝐝 From Huré–Pham–Warin (2019), solves semilinear PDEs by reformulating them as backward stochastic differential equations. Trains one NN per time step (forward simulation) – robust for pricing American/early exercise options. 🧠 Widely used in quantitative finance for high-dimensional PDEs. ➡ 𝐃𝐆𝐌 (Sirignano & Spiliopoulos, 2018) Solves time-dependent PDEs by training a single NN on space–time, using Galerkin-inspired residual loss. No time stepping needed – clean and efficient on smooth solutions. ➡ 𝐓𝐃𝐆𝐅 (Papapantoleon & Rou, 2024) Reformulates PDE solving as energy minimization per time step. Very stable, compatible with higher-order schemes, great for option pricing. — ⏳ 𝗦𝘁𝗶𝗹𝗹 𝘄𝗮𝗶𝘁𝗶𝗻𝗴 𝗳𝗼𝗿 𝗮 𝗴𝗲𝗻𝗲𝗿𝗮𝗹-𝗽𝘂𝗿𝗽𝗼𝘀𝗲 𝗺𝗲𝘁𝗵𝗼𝗱? Every architecture has trade-offs: PINNs are versatile, BSDEs are robust, DGM is elegant, TDGF is precise. This paper provides practical guidelines to tune and compare solvers, beyond just theory. 🔗 https://lnkd.in/esEyvtjy
11
-
Wen Gong
631 followers
The following 2 papers were accepted at arXiv last week: - Benchmarking Cross-Lingual Semantic Alignment in Multilingual Embeddings ( https://lnkd.in/eNte3aQW ) - Geometric Patterns of Meaning: A PHATE Manifold Analysis of Multi-lingual Embeddings ( https://lnkd.in/ew2AFY-d )
22
-
Mike Carvill
The MathWorks • 4K followers
𝗙𝗿𝗼𝗺 𝗠𝗼𝗻𝘁𝗵𝘀 𝗶𝗻𝘁𝗼 𝗢𝘃𝗲𝗿𝗻𝗶𝗴𝗵𝘁: 𝗔𝗰𝗰𝗲𝗹𝗲𝗿𝗮𝘁𝗶𝗻𝗴 𝗠𝗮𝗰𝗿𝗼 𝗙𝗼𝗿𝗲𝗰𝗮𝘀𝘁𝗶𝗻𝗴 𝘄𝗶𝘁𝗵 𝗠𝗔𝗧𝗟𝗔𝗕 How can economists generate reliable confidence bands for nonlinear DSGE models without waiting weeks for results? At the MathWorks Finance Conference, Kadir Tanyeri (International Monetary Fund) showed how a MATLAB‑based workflow accelerates high‑dimensional macroeconomic forecasting—cutting simulations from 100,000+ to about 3,600. Using a fully automated, nonlinear forecasting pipeline, the workflow integrates model solution, data preparation, filtering, and large‑scale simulation in one environment. 🔍 𝗛𝗲𝗿𝗲'𝘀 𝘄𝗵𝗮𝘁 𝗺𝗮𝗸𝗲𝘀 𝗶𝘁 𝘄𝗼𝗿𝗸 🔹 Nonlinear inflation dynamics with asymmetric output‑gap effects 🔹 Automated data transformation and baseline forecast generation 🔹 Latin hypercube sampling for faster convergence in a 1,680‑dimensional shock space 🔹 Distributed computing on a 128‑worker cluster to scale simulations efficiently 🔹 Single‑function execution of full nonlinear confidence‑band computation ℹ️ 𝗪𝗵𝘆 𝗶𝘁 𝗺𝗮𝘁𝘁𝗲𝗿𝘀 Macroeconomic projections—GDP, inflation, and unemployment—are only useful when paired with quantified uncertainty. Confidence bands support: 🔹 Risk‑aware policy decisions 🔹 Recession and deflation probability estimation 🔹 Clear communication of upside/downside risks As nonlinearity and dimensionality grow, traditional Monte Carlo approaches become computationally prohibitive. MATLAB provides a unified, reproducible, and scalable environment for modern macroeconomic forecasting. Have you used Latin hypercube sampling or distributed computing in your forecasting work? 🔗 𝗥𝗲𝗮𝗱 𝘁𝗵𝗲 𝗳𝘂𝗹𝗹 𝗯𝗿𝗲𝗮𝗸𝗱𝗼𝘄𝗻 𝗵𝗲𝗿𝗲: https://spr.ly/6048hCT3p #Macroeconomics #Forecasting #DSGE #ParallelComputing #Econometrics #RiskManagement
1
-
Simon Villani, PhD
ANZ • 25K followers
Stop treating your favourite LLM like a single source of truth. Andrej Karpathy just dropped “LLM Council” – a simple local app where multiple models answer your question independently, then review and rank each other’s answers, and finally a “chairman” model synthesises the final response via OpenRouter. This matters because it: - Assumes models are fallible and need cross checking - Builds critique and review into the workflow - Focuses your effort on orchestration, not hero-worshipping one provider If you are a student or early-career dev, do not build yet another single-model chat UI. Fork this pattern and ship your own “council” on a real problem in your domain. Repo: https://lnkd.in/ghj4hjJ2
28
5 Comments -
GoldenSource
26K followers
What do Black-Scholes, NLP, and yield curve modeling have in common? More than you'd think. In his last July blog, Charlie Browne traces the surprising parallels between quantitative finance and AI, from Ito’s calculus to generative models. Why it matters: Whether you're pricing options or optimizing neural nets, both disciplines rely on probabilistic thinking, curve fitting, and risk minimization. And both are driving innovation in how we model the future. Read now: https://lnkd.in/ehGTrSZp
9
-
Greig Cowan
NatWest Group • 5K followers
Fantastic work by the team here. So many interesting questions are being raised by the adoption of LLMs and it is only through performing novel research such as this, coupled with real-world applications and experimented of using them that we will be able to develop robust controls and guardrails that allow us to use AI safely. #responsibleAI
19
-
Craig Whiting
RLS Search • 27K followers
Where are Funds putting their AI-Money? Based on the current roles we're focused on, there's large investment in LLM integration for Quantitative Workflows, Time-Series Forecasting/Simulation and Reinforcement Learning/Agentic Systems for Research Specific focus on NLP within unstructured data and RL environment generation around time-series modelling, backtest, simulation and rebalancing, aswell large-scale research and analytics capability. Roles are a healthy mix of Quantitative, Software, Data and Product, all at Senior IC and all with a clear orientation toward AI/ML commercial experience. All firms are looking for hands-on individuals who can blend an exec presence with a sleeves-up approach. Vast majority of hiring sentiment is focused on ICs, not leadership. Most of the leadership roles are filled. Larger firms still favour financial backgrounds and will wait for lengthy NCs to secure the right talent. Smaller firms are open to buy-side to buy-side shift, but are more aggressively looking at MedTech, Pharma, FAANG or Tech StartUps as their source of exceptional talent. Compensation is through the roof, the battle for talent is intensifying. All very much to play for.
14
7 Comments -
Hua Hua
Capital One • 815 followers
🚀 New Story on Medium: Hands-On with Hugging Face LLM Course – GLUE Text Classification I’ve just published a new article where I walk through fine-tuning BERT for the GLUE benchmark as part of my Hugging Face LLM learning journey. 🔹 What’s inside: Why GLUE is a gold-standard benchmark for NLP models How to fine-tune BERT for paraphrase detection (MRPC task) Lessons learned from combining Hugging Face hands-on work with Stanford CS224n theory 💡 Whether you’re starting in NLP or looking to refine your LLM skills, this project will give you a practical, end-to-end path — from dataset to evaluation. #NLP #MachineLearning #HuggingFace #BERT #DeepLearning #LLMs #AIResearch #CS224n #AI #ArtificialIntelligence 📅 Coming Next Week: Building a Retrieval-Augmented Generation (RAG) Pipeline — where we combine LLMs with external knowledge sources for more accurate, context-aware responses. Using LangChain with Hugging Face models Implementing document retrieval + embedding search Deploying a working RAG app in Hugging Face Spaces Stay tuned! 📖 Read here: [https://lnkd.in/e4nuPUTw]
10
-
Luca Massaron
illimity • 18K followers
Large language models dominate AI headlines, but in real engineering environments, something else is quietly happening: the rise of small, specialized models. In this Medium article, we take Gemma 3 1B-IT, a compact instruction‑tuned SLM, and transform it into a domain‑aware financial sentiment analyst. Instead of teaching the model only to classify text as positive, neutral, or negative, we train it to explain its reasoning, an essential skill in financial applications where transparency matters. #Gemma3 #Gemmaverse #AI #LLM #FineTuning #SentimentAnalysis #NLP
26
-
Marcus Tassler
NORD/LB • 254 followers
Interesting read. As Claude is giving answers Anthropic has been tinkering with the activations of its Neural Net in 50% of the cases and the model is to report for all test and control cases if it is noticing any concept injection into its thought processes. The latest Claude Opus model in particular reported a manipulation of its internal states for considerably more test cases with actual external intervention than reporting false positives for control cases without external intervention. The difference between the correct identification of concept injections and false positives was 20% and statistically significant: „Taken together, our experiments suggest that models possess some genuine capacity to monitor and control their own internal states. This doesn’t mean they’re able to do so all the time, or reliably. In fact, most of the time models fail to demonstrate introspection—they’re either unaware of their internal states or unable to report on them coherently. But the pattern of results indicates that, when conditions are right, models can recognize the contents of their own representations. In addition, there are some signs that this capability may increase in future, more powerful models (given that the most capable models we tested, Opus 4 and 4.1, performed the best in our experiments).“
2
-
Ryan Chaves, Ph.D.
Mollie • 1K followers
"Advanced Data Science for Credit Risk Modeling: The Good, The Bad, The Defaulted" Watch our ML Engineer Marta Barigozzi's presentation from the recent #PyData Milan event. She talks about how the ML Craft at Mollie carefully mitigates risk and ensures we make the best possible predictions for what's called probability of default. These ML models are the secret sauce for how we are able to provide better financing through Mollie Capital to help our customers grow their businesses, while we all stay healthy financially. Marta shares pro-tips for model developers, like leveraging Partial Dependency Plots (PDPs) and monotonicity constraints to reduce overfit and ensure models perform as expected in production. She also speaks about model explainability: how a combination of smart SHAP-based visualizations and AI can help both ML Engineers and credit risk colleagues understand what drives each model prediction. No trade secrets here: just a thoughtful mix of sometimes forgotten old-school modeling techniques, open-source software like sklearn, shap, and feature-engine, and, of course, AI. https://lnkd.in/eZsTZDDR
32
1 Comment
Explore collaborative articles
We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
Explore More