I am pursuing a PhD under the supervision of Stanislas Dehaene at PSL/Collège de France and NeuroSpin, CEA. I’m interested in humans’ striking ability to manipulate highly abstract structures, be it language, mathematics or music. My work focuses on the perception of geometry, seeking traces of the ability for abstraction in a domain attested to be extremely old: homo erectus already carved abstract geometrical patterns half a million years ago, while other non-human primates seem unable to produce such shapes.
My work relies on experimental psychology to answer specific questions: are certain shapes processed faster and perceived more accurately than others? Even when matched for low-level perceptual features? What characterizes such shapes, and why? We’ve run comparative experiments with French adults at a large scale, together with behavioral data from preschoolers, uneducated adults, and neural networks. This work is moving toward incorporating neuroscience methodologies to answer new questions: EEG in babies to get even more naive participants, and fMRI & MEG in adults to look for perception-independent representations of geometrical shapes.
I also have strong interest in the topics below:
- (i) Either Jane is English and Sue is Dutch, or
Claire is English.
(ii) Jane is English.
(Prompt) Does it follow that Sue is Dutch?
No, but 70% say yes!
Human reasoning and its (in)dependence from natural language: to what extent do systematic fallacies in reasoning hinge on the fact the meaning of sentences is enriched through pragmatics? I argue that we can understand fallacies as instances of humans engaging in a different game. Building on the theoretical work of the Erotetic Theory of Reasoning, and bridging the gap with Bayesian theoretic literature, I argue that humans are not trying to “maximize what’s true” (either when speaking or reasoning) but rather “maximize what’s informative and useful”. This worked is carried out in collaboration with Salvador Mascarenhas at ENS, Institut Jean-Nicod, LINGUAE team
Learning a generative model over LOGO/Turtle graphics programs. Shown are renderers of randomly generated programs from the learned prior.
Human sequence processing: it is easy to find long sequences of few elements that are easy to remember (think a, b, c, b, a, b, c, b,…). But characterizing what makes such sequences easy proves to be very hard. This has links with geometry when such sequences unfold on a plane to form a geometrical shape, in which case the nature of the rules that humans are able to use is informative about the internal representation of the unfolding sequence.
- Program induction: the nature of the structured representation of many of the high-level concepts mentioned above (geometry, reasoning, language) makes it tempting to model them as computer programs. Then one has to wonder: how does one go from world perception and stimuli to abstract, internal representation? Is it always possible, and how can it be done efficiently? The subdomain of program induction I am interested in tackles these very questions: starting with a specific idea of what the structure looks like, and a bunch of “real world tasks” (i.e. examples of the input-output produced by the programs, but not the programs themselves), can we make proof-of-concept algorithms that find accurate representations? I was worked on this together with Kevin Ellis and many great people at MIT’s CBMM’s CoCoSci group, piloted by Josh Tenenbaum.
See dedicated page, which I used when TAing logic to Philosophy and Cognitive science ENS students in their first year of Master Degree.
After studying math, physics and computer science in prep. classes in France, I entered École Normale Supérieure (ENS) de Cachan in Computer Science where I finished my License (BA) and my first year of Master. During my masters I spent six months in Oxford, UK doing theoretical computer science under the supervision of Luke Ong, working on the semantics of various λ-calculi.
I then took a gap year sailing
A dinner on the island of Ligia, Greece during my gap year. Boat in the background, Félicien Comtat in the foreground.
and decided to focus on cognitive neuroscience: I applied for the CogMaster and worked with Stanislas Dehaene on geometrical sequences. Between that year and my PhD I spent six months at École Normale Supérieure working under the supervision of Salvador Mascarenhas on the links between reasoning and language, as well as six months at MIT under the supervision of Josh Tenenbaum working on program induction and more specifically applying it in the domain of geometry.
I have been coadvising, together with my own supervisor, the M2 internship of Maxime Cauté. He is exploring cross-modal representation of sequences of parametrized complexity using language-of-thought models.
- A language of thought for the mental
representation of geometric shapes
Mathias Sablé-Meyer, Kevin Ellis, Joshua Tenenbaum, Stanislas Dehaene (under review: preprint available on PsyArXiV and data available on OSF)
- Question-answer dynamics and confirmation
theory in reasoning by representativeness.
Mathias Sablé-Meyer, Janek Guerrini. Salvador Mascarenhas (under review: preprint available on PsyArXiV and data available on OSF)
- Analyzing the misperception of exponential
growth in graphs.
Lorenzo Ciccione and Mathias Sablé-Meyer, Stanislas Dehaene
Published in Cognition
- Indirect illusory inferences from disjunction:
a new bridge between deductive inference and
Mathias Sablé-Meyer, Salvador Mascarenhas,
Published in Review of Philosophy and Psychology, PDF available to view here. Associated data and scripts, tools for the study of mSentential
A signature of human uniqueness in the perception of geometric shapes
Mathias Sablé-Meyer, Joel Fagot, Serge Caparos, Timo van Kerkoerle, Marie Amalric, Stanislas Dehaene,
Published in PNAS, PDF available here
- DreamCoder: bootstrapping inductive program
synthesis with wake-sleep library learning
Kevin Ellis, Catherine Wong, Maxwell Nye, Mathias Sablé-Meyer, Lucas Morales, Luke Hewitt, Luc Cary, Armando Solar-Lezama, Joshua B Tenenbaum
Published in PLDI 2021
- Library Learning for Neurally-Guided Bayesian
Kevin Ellis, Lucas Morales, Mathias Sablé-Meyer, Armando Solar-Lezama, Joshua B. Tenenbaum.
NIPS 2018. Spotlight.
Talks & Seminars
- Invited speaker at the CoLaLa, invited by Steven Piantadosi: “A language of thought for the mental representation of geometric shapes”, 2022
- Invited speaker at the Department of Psychology and Neuroscience, Temple University, invited by Kathryn A. Hirsh-Pasek and Nora Newcomb: “A language of thought for the mental representation of geometric shapes”, 2022
- Invited speaker at the McDonnell plenary workshop 2022, “A language of thought for the mental representation of geometric shapes”, 2022
- Invited speaker at the Brain/AI at Facebook AI Research (FAIR), invited by Jean-Rémi KING: “Sensitivity to geometric shape regularity in humans and baboons: A putative signature of human singularity”, 2021
- Chairman for the Fondation Les Treilles “Cognitive maps in infants: Initial state and development” 2021
- Invited member of the seminar “Music, Brain and Education”, organised by Oubradou/Collège de France, 2020
- Invited speaker at the LINGUAE Seminar: “The laws of mental geometry in human and non-human primates”, 2019
- FYSSEN seminar “Pillars of cognitive development in mathematics”, 2019
- Invited young researcher at Centre l’Oubradou, “Where Art, Science & Education connect”, 2018
- Joint talk with Kevin Ellis: “Dream-Coder: Bootstrapping Domain-Specific Languages for Neurally-Guided Bayesian Program Learning”, at CogSci 2018 workshop on program induction
- LPPRD seminar, joint talk with Salvador Mascarenhas, invited by Philip Koralis, 2018, handout.
- The Experimental Philosophy Group, 2017, handout
- CogSci 2021 : Sensitivity to geometric shape
regularity in humans and baboons
Mathias Sablé-Meyer, Joel Fagot, Serge Caparos, Timo van Kerkoerle, Marie Amalric, Stanislas Dehaene,
Poster available here; more information available on the summary page of the poster
- CogSci 2019 : Mathias Sablé-Meyer, Salvador Mascarenhas. “Assessing the role of matching bias in reasoning with disjunctions”
Structural compression of visual input?
- Visual Sequence Primitives in Humans, 2017, internship report from my work in the NeuroSyntax team of the UniCog lab in NeuroSpin, under the supervision of Stanislas Dehaene.
- Toward an unstaging translation for an environment classifier based multi-staged language, 2015, internship report from work supervised by Luke Ong in Oxford’s Computer Science Department.
- Algorithmes de recommandation séquentielle, 2014, internship report with INRIA Lille’s team SequeL, supervised by Romaric Gaudel.
- I had fun with Jacquin’s algorithm for fractal compression of images in 2013, you can find some of that work on the algorith on my former webpage
General public communications
- Two articles from my PhD were featured in a great piece written by Siobhan Roberts for the NYT
- About the PNAS article A signature of human
uniqueness in the perception of geometric shapes
- A communication from the AFP (French AP), shared many times including here: Les symboles de la géométrie signent peut-être la singularité de l’être humain (also translated in several languages, not listed here)
- A communication from Collège de France: Les humains sont dotés d’un sens unique de la géométrie
- Several communications from CEA and institut Joliot: Les humains sont dotés d’un sens unique de la géométrie
- DreamCoder has been the subject of a video from Yannic Kilcher where he reads through the articles and explains key concepts and important decision points.
I have been directly contacted and have provided reviews for the following journal and conferences:
- 2022, two reviews for the conference “CogSci 2022”
- 2021, one review for an article in the journal “Cognitive Science”
I’m an avid climber, a competent sailor, an adequate handyman but a terrible mechanic. Occasionally I get hooked into tinkering with various programming languages, neural networks, etc., and when in luck I write about it. Selected examples here:
The documentation for a small project I realised during my time as an intern at NeuroSpin — reverse engineer a food distribution system to interface it with a computer through USB.
- Interfacing OCaml and Rust or playing with FFIs and Ocaml/Rust
- Messing with DeepL, just for fun.
- Messing with GPT-2 117M, again.
- A course from CogMaster was validated through a mini thesis. It’s about Mathematical Intuitions and Monsters (in French).
- More of the background of this page?
- My name without an accent is “Mathias Sable-Meyer”. I have been inconsistent in my use of handles in the past, and used: “mathsm” at MIT, “mathias-sm” on github, “@(M)SableMeyer” on Twitter, “msm” whenever it’s free, “msableme” when a university choses for me.