Joey David

Joey David

NLP/ML Student Researcher

I'm a Student Researcher focused on reasoning and self-improvement in LLMs and Machine Learning in general.
Aside from completing my Master's Degree in AI/ML at the Paris Sciences & Lettres University (supervised by ENS, Mines and Dauphine), I'm currently working on semi-supervised learning for medical imaging with INSA Lyon (DISP Lab). After this, I'm looking to continue in applied research in LLMs.
I sometimes build open-source tools and agents (which you can find below), for self-teaching and for fun.

Research

  • Semi-Supervised Pipeline for Etiological Diagnosis of Uveitis

    Ongoing

    J. David, T. Wang, R. Jacquot.

    INSA Lyon, DISP Laboratory • 2025

    Developed a semi-supervised ML pipeline incorporating Multi-Task Siamese Networks (MTSN), One Vote Veto (OVV) dataset extension, and Graph Attentional Convolutional Neural Networks (GACNN) for automated diagnosis of sight-threatening ocular inflammation.

    Preview

Selected Projects

  • Agentica

    Modular agentic framework for automation, learning, and research. Current agents include a research agent, an email assistant for categorization and others. Implementation of Multi-Agent Systems (MAS) through dynamic agent generation and management for complex task execution.

  • GPT-2 Implementation

    Multi-head attention mechanism and transformer architecture, trained on the tinyshakespeare dataset. Later added extra features such as automated finetuning on messaging datasets for simulation of personas.

  • Anomaly Detection System

    Fine-tuned PatchCore model for industrial soldering defect detection using anomalib framework.

  • C++ From-Scratch Neural Networks

    C++ and numpy implementation of feedforward neural networks for advanced character recognition, music classification and others.

Education

  • 2025-26

    M.Sc. Artificial Intelligence, Systems and Data

    ENS Ulm, Université Paris-Dauphine, Mines ParisTech - PSL

  • 2024-25

    M.Sc. Computer Science (M1)

    Université Claude Bernard Lyon 1

  • 2020-24

    B.Sc. Computer Science

    Université Claude Bernard Lyon 1

    Exchange Student at Oregon State University (2023-24)

Professional Experience

  • 2025

    Graduate Student Researcher

    INSA Lyon, DISP Laboratory - Semi-supervised ML for Medical Imaging

  • 2024

    Software Engineering Intern

    Sopra Steria - NLP and Full-Stack Development

Misc.

  • ChatBrain

    Conversation analysis platform with custom YOLO detector for chat interface recognition and OCR transcription.

  • Playsorter

    Spotify playlist organization tool using Jaccard similarity metrics and ranking algorithms.

  • py-scripts

    Collection of Python scripts I've been creating for various tasks and automation. You may find something you like there!

Thoughts & Rants

The Goal Always Was ASI

This is still being written, reach out to discuss it!

The Brain as an Ideal: Parallels and Shortcomings

I. The networks we make, the networks we are

Despite their seemingly endlessly increasing sizes, human-engineered neural-network based systems remain easy to modelize, track and “explain”.

Though we can’t track their every weight at every time in a perfectly deterministic way, we can shape them to the best (or at least a local best) of a loss landscape we get to define.

It’s now almost undisputed by anyone with a technical understanding of LLMs that they are simply very proficient pattern-matching machines; even their finetuning can be seen as simply bending the patterns they produce.

On the other hand, we don't understand our own brains nearly as well as the networks we’re producing.

We more or less know that specific areas of the brain specialize in specific tasks and responsibilities, we know that their building block is the neuron, we know there are different speeds of reasoning and types of actions but that’s an entirely superior tier of ignorance to not knowing exactly which attention head attends to what.

We don’t understand them because we didn’t build them from scratch like we did transformers, and thus, we can’t truly tell if they perform something fundamentally above extremely proficient and wide pattern matching.

II. Two of a kind?

Regardless of how little we know about our own brains, we can’t help but observe so many similarities between the systems we build in search of human-like (or better-than-human) intelligence, and the systems that have already achieved human intelligence :

>Reinforcement learning, gradient descent, and evolution are three names for the same fundamental process.
Neural nets learn behavior by watching thousands of examples, until the right connections are made between their components to functionally reproduce that pattern, through reward systems. In the same way, the best way to learn as a human is to practice, to do, literally reinforcing the neural circuits involved in that action until they’ve mastered the behavior.
Our bodies, which are so well crafted that many theists use them as an argument for the existence of a supernatural creator, are but the fruit of the same process once again.


Diagram contrasting brain inputs and outputs

At first, it’s hard to conceive of the human brain as a unidirectional neural network, with a fixed number of outputs. If its inputs are fixed, surely they are of an immense number, and of a non-discrete form - the images in our eyes aren’t exactly arrays of pixels, our sense of taste is hard to express as the span of a finite basis, same goes for our proprioception, hearing, sense of touch, etc, doesn’t it?

And yet, the brain receives external information in discrete spikes along finite nerves, predetermined and of a given throughput. What about the brain’s “outputs”? Intuition would tempt us to say that it is more than simply motor commands to the body (again, transmitted by specific nervous channels), to treat thoughts as products in their own right. But unless a thought is transduced into something more than neural activity, is there any route other than motor control by which it becomes real? [sure, you could mention brain-computer-interfaces, but that’s the same as sampling inner activations - we do that with llms too.]


But not all apparent differences can be overcome. Transformers and human brains radically differ in self-modification. This is probably what allows us to learn complex workflows, methodologies and behaviors in no more than a few exposures, when neural nets need thousands, millions or billions of examples.

Where transformers are fundamentally discrete, operating one token at a time, one iteration of attention before the other, the brain is time-continous in its operation, and orchestrates itself in a web of nuance, parallel neural circuits and hierarchical speeds that is hard to rival.

The century or so that has occurred since the discovery of the perceptron is nowhere near enough to truly rival 4 billion years of iteration and refinement. And yet - multi-scale models are beginning to see the day. The hierarchical reasoning model, coupling two LLMs of different scales to model thinking fast and slow, has been impressive on its performance (albeit on very restricted and specific tasks). As we work on size and computing efficiency, where we currently lag so far behind the human brain it can seem inconceivable to ever catch up, maybe all of the fundamentally “different” and “superior” abilities we take pride in will manifest themselves in intelligent, artificial systems, one by one.

(What) Will It Want?

This is still being written, reach out to discuss it!