Skip to content
View Whojo's full-sized avatar

Block or report Whojo

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Whojo/README.md

Hi 👋, I'm a PhD student in Deep Reinforcement Learning & Robotics

TLDR;

  • ✨ Open-Source Contributor of AlphaZero.jl: MCTS on CUDA for 13x speed improvement
  • 🔬 Research on Distributed Reinforcement Learning in Robotics & Self-Supervised Natural Language Processing @ LRDE & CEA (France)
  • 💻 Double CS Master and Engineering degree @ Epita & Sorbonne in Computer Vision
  • 🔥 Interested in Reinforcement Learning, Computer Vision & High Performance Computing
  • 🚴 Love Triathlon, Bouldering & Board Games
  • 🇬🇧 Speaking English & French

🛠️ My Skills

PyTorch Sklearn Tensorflow Pandas NumPy Slurm Docker Git Jupyter

🎓 My Background

I was introduced to the world of AI and research during my first internship in my school's AI laboratory, where I used Machine Learning (sklearn) methods to improve the detection of malicious traffic in network cores. My contribution led to a 50% reduction in errors compared to the previous model.

Recently, I contributed to the Open-Source community through the Google Summer of Code program. AlphaZero, an algorithm developed by Google Deepmind, has made significant advancements in the fields of chess, Go, and Shogi through its Reinforcement Learning approach. To make this algorithm more accessible, AlphaZero.jl has been created as a powerful Open-Source implementation of this algorithm in Julia. I parallelized the environment exploration using CUDA, which resulted in a 13x speedup compared to the previous implementation, as described in this accessible Medium article.

During my studies in France at Epita and my double master's degree at the Sorbonne, I specialized in Machine Learning and Deep Learning methods for Computer Vision, extensively experimenting with Tensorflow and PyTorch Framework. Meanwhile, I joined Epita's Research and Development Laboratory, where I am researching and developing tools to automatically detect and correct texts extracted from old documents. My latest project involves a supervised approach using a sequence-to-sequence model (BERT), which we plan to publish a paper on in May. This experience has allowed me to deepen my research background and gain experience with RNNs and Transformers in the Natural Language Processing context. To explore further details about this project, please feel free to consult my final report, available here.

Further on the path of computational resources optimization, I am currently undertaking a research internship at the French Alternative Energies and Atomic Energy Commission. In this position, I am responsible for distributing the computing load for training a state-based Reinforcement Learning algorithm, which is being used to control a multi-task robotic arm, by utilizing Jax (\cite{jax}). The goal of this project is to reduce the current four-day training time down to just one hour, resulting in a significant 100x speedup. This acceleration will greatly enhance our ability to iterate through experiments and generate meaningful results, in the hope to reproduce results of SayCan with fewer resources.

In addition to my academic pursuits, I am also involved in several side-projects, including the President of my School's Chess club, Co-Organizer of an AI Safety discussion group, Head of my school's Student Office Events, and participation in Reinforcement Learning side-courses.

I value teamwork and effective communication within a team. I was able to practice the latter when I co-organized a discussion group on AI safety for nine months and was president of Epita's chess club for three years. I am autonomous, hardworking, and passionate. By actively contributing to developing these technologies, I hope to become one of the critical actors in the following decades.

📫 How to reach me:

Pinned Loading

  1. AlphaZero.jl Public

    Forked from jonathan-laurent/AlphaZero.jl

    A generic, simple and fast implementation of Deepmind's AlphaZero algorithm.

    Julia 4

  2. Optical-Character-Recognition Public

    Deep Neural Network coded from scratch in C

    C

  3. Gradient_Descent_Comparisons Public

    Benchmark on different Gradient Descent algorithms

    Jupyter Notebook 1 1

  4. Bag_of_Visual_Words_search_engine Public

    Jupyter Notebook

  5. Brain_Segmentation Public

    Jupyter Notebook

  6. Raytracer Public

    A CPP Raytracer

    C++

17 contributions in the last year

Contribution Graph
Day of Week April May June July August September October November December January February March April
Sunday
Monday
Tuesday
Wednesday
Thursday
Friday
Saturday
Less
No contributions.
Low contributions.
Medium-low contributions.
Medium-high contributions.
High contributions.
More

Contribution activity

April 2025

Whojo has no activity yet for this period.
Loading