<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://adntaha.github.io/feed.xml" rel="self" type="application/atom+xml" /><link href="https://adntaha.github.io/" rel="alternate" type="text/html" /><updated>2025-10-14T00:56:00-04:00</updated><id>https://adntaha.github.io/feed.xml</id><title type="html">Aidan Taha</title><subtitle>The personal website of Aidan Taha, a maker, a novice writer, and an aspiring AI Safety researcher.</subtitle><author><name>Aidan Taha</name></author><entry><title type="html">CD Player Kidnapping Saga</title><link href="https://adntaha.github.io/blog/cd-player-kidnapping-saga/" rel="alternate" type="text/html" title="CD Player Kidnapping Saga" /><published>2025-09-21T14:30:20-04:00</published><updated>2025-09-21T14:30:20-04:00</updated><id>https://adntaha.github.io/blog/cd-player-kidnapping-saga</id><content type="html" xml:base="https://adntaha.github.io/blog/cd-player-kidnapping-saga/"><![CDATA[<p>I helped a friend with fixing an old, 1990s-era<sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup>, CD player that belonged to his grandparents. But before we get into that, we need to talk about how I was convinced into helping him<sup id="fnref:2" role="doc-noteref"><a href="#fn:2" class="footnote" rel="footnote">2</a></sup>. It all started quite ominously:</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/kidnap-dms.png" alt="Instagram DM screenshot, from my friend on Friday at 1:34PM : Mr. Taha, Je te kidnap après anglais, Tu as été prévenu. My response is &quot;pardon.&quot; Lmao" /></p>

<p><em>Translation: “Mr. Taha, I’m kidnapping you after English, you’ve been warned”, and my response was “what.” For context, our English literature class had just started when I received this DM from him. It was a few confusing seconds before I realized what he meant.</em></p>

<p>The more helpful piece of context came from another friend, let’s call him Nicolas, which ran the community bike repair shop whose workspace we were about to borrow for about 5 hours.
<img src="/assets/images/cd-player-kidnapping-saga/helpful-dms.png" alt="Instagram DM screenshot, from another friend on Friday at 1:34PM : T'as un cours rn?, I respond &quot;oui&quot;, he responds at 1:42 PM, &quot;Maidadouille, Hadrien a besoin d'aide&quot;, along with a picture of the CD player open on a table." /></p>

<p><em>Translation: At 1:27 PM on Friday, “Are you in class right now?”, “Yes”, and at 1:42 PM, “Midadouille, Hadrien needs help.” [Friend #1] is the friend who sent those ominous messages. I have no idea what “Midadouille” is or where it came from.</em></p>

<hr />

<p>Skipping to the part where we start working on it, we unscrew it, and find that everything is electronic except the part that pops out and reads the CD, which is connected to the rest using a bunch of connectors.</p>

<p>Using the tools at our disposal, we spend at least 30 minutes unplugging a few of these <a href="https://en.wikipedia.org/wiki/JST_connector#/media/File:Balancer_Buchse_XH.JPG">JST-style connectors</a>, most of which were much more annoying because they had two bits that locked into their casing which we needed to kind of push out on the casing a bit to get them out.</p>

<p>We then took it out (which was pretty straightforward because we had access to the owner’s manual, which had so. much. information.)</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/friend-removing-cd-tray.png" alt="My friend, holding the CD tray in his left hand and the whole assembly with his right hand, trying to remove the former from the latter." /></p>

<p><em>Taking off the front panel to access the CD reader, which my friend Hadrien is holding with his left hand.</em></p>

<p>The CD reader could be split into two parts, which exposed two rotating shafts, essentially two wheels, a lot more plastic, a spring which served to push up the CD when it was ejected so that it would be easier to take out<sup id="fnref:4" role="doc-noteref"><a href="#fn:4" class="footnote" rel="footnote">3</a></sup>.</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/spring-system.png" alt="A spring connected to a gear and the plastic panel which it's installed on. It's protected by some tubing." /></p>

<p><em>The spring system that pushed up the CD tray. We actually thought for a bit that it was a part of the problem, but turns out it was perfectly fine and working as intended.</em></p>

<p>It’s also important to note that we had a weirdly shaped piece<sup id="fnref:3" role="doc-noteref"><a href="#fn:3" class="footnote" rel="footnote">4</a></sup> and a bunch of strands of rubber loose inside the main case. We didn’t know what their use was. In fact, it took a lot of trial and error to get the weirdly shaped piece to fit somewhere, never mind finding where it was actually supposed to go.</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/rubber-band-fragment.png" alt="A picture of a piece of rubber that's not a whole circle but only a part of one." /></p>

<p><em>One of the rubber band pieces in question.</em></p>

<p>There were also two plastic wheels on the other piece of plastic. It took us embarrassingly long to figure out that the rubber band was supposed to link between them, but we eventually did. However, issue #1, we did not have any rubber bands that fit the specs. By using a piece of tape that I wrapped around them, I figured out that we needed something with a 10cm circumference and a 1-1.5mm diameter. And unfortunately they don’t have small specific rubber bands at a community bike repair station, I even went up to ask and the answer was still a no. :(</p>

<p>So I got crafting. I tried shrinking a really close rubber band that we somehow found at the corner of a desk by melting it with a hot glue gun (it didn’t work), using electrical tape (green in the image below), masking tape, even making one out of hot glue (this actually almost worked).</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/new-rubber-band-attempts.png" alt="A lot of wires and circles made of electrical tape, masking tape, hot glue, some pieces of rubber band. It's not a very good looking desk, very disorganised, you're lucky that you have trouble seeing it, trust me." /></p>

<p><em>A bunch of our attempted solutions. The masking tape didn’t have enough friction, the hot glue band had deformities which made it jump out, but the electrical tape one was the closest to working.</em></p>

<p>After 7-8 attempts with the electrical tape, changing its thickness, length, and how I joined the two sides, we ended up with something that worked! Or at least well enough. I wouldn’t be able to guarantee more than 20 spins because its adhesive wasn’t strong enough, but it was a working PoC until we ordered an actual rubber band.</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/munched-up-rubber-band-attempt.png" alt="A rubber band made of electrical tape that looks like it's been sandwhiched between two gears while they were running." /></p>

<p><em>One of our attempts was munched up by a gear that was right next to the belt and pulley system where we installed it.</em></p>

<p><img src="/assets/images/cd-player-kidnapping-saga/perfect-rubber-band-attempt.png" alt="An rubber loop made from green electrical tape that looks so much like a regular green rubber band." /></p>

<p><em>The holy grail, an almost perfect looking *fake* rubber band.</em></p>

<p>At the same time, we had to figure out where the weird looking piece went. And after a ton of trial and errors, we found where it fit, but we inserted it wrongly which ended up snapping off a piece of plastic. Once we installed it correctly we were able to use hot glue to mitigate that issue.</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/weird-looking-piece.png" alt="I seriously don't know how to describe this piece. It turns rotation into translation, and it's got a crescent shape. It's held on by hot glue at its center." /></p>

<p><em>Weird looking piece in place (the one in beige), with some hot glue keeping it from falling off ever again.</em></p>

<p>After fixing these two items, even if some things on the CD reader weren’t working as expected (in fact, the CD tray raising mechanism was working in reverse), we decided to try it out, and the CD tray would pop out and raise the tray correctly! That said, it didn’t want to read the disk we were putting into it, so back to trial and error. It turns out that the laser disk reading head was stuck, so it just took a bit of wiggling to get it moving, and 🎉. We didn’t see any other issues, and the numbers on the LCD made it seem like the audio was playing.</p>

<p>So at least it seemed to work, because to actually hear the audio we needed to connect it to a pre-amp and then an amp, which we didn’t have on the spot, so my friend took it back home, and lo and behold.</p>

<p><img src="/assets/images/cd-player-kidnapping-saga/success-dm.png" alt="An Instagram story my friend sent me of his sound system, with the caption &quot;@adn.taha IL FONCTIONNE&quot; in the middle." /></p>

<p><em>Translation: “@adn.taha IT WORKS.” The CD player is the item at the bottom of the stack of audio equipment, the one with the LCD display.</em></p>

<p>All that’s left is to order an actual 1mm diameter 10cm circumference rubber loop, so if you know where to get something like that, please contact me!</p>

<div class="footnotes" role="doc-endnotes">
  <ol>
    <li id="fn:1" role="doc-endnote">
      <p>According to him. According to this <a href="https://www.hilberink.nl/codehans/datumlux.htm">source that I found</a>, it was made in February of 1989, which is basically the 1990s, so he’s right and we should never doubt our friends. It’s a Luxman D-105U. <a href="#fnref:1" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:2" role="doc-endnote">
      <p>Spoiler alert, it didn’t take much convincing. <a href="#fnref:2" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:4" role="doc-endnote">
      <p>The type of feature built into luxury items that’s a pain to fix. Luxury cars also do this, which is why mechanics dislike working on them. <a href="#fnref:4" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:3" role="doc-endnote">
      <p>If you think about it, all mechanical pieces are “weirdly shaped” because they’re made for a specific purpose, which means that none of it is really “weird” at all. <a href="#fnref:3" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
  </ol>
</div>]]></content><author><name>Aidan Taha</name></author><category term="The Workshop" /><category term="Analog Systems" /><category term="Reverse Engineering" /><summary type="html"><![CDATA[How I got convinced to help fix my friend's grandparents' CD player.]]></summary></entry><entry><title type="html">What’s AI Safety even about?</title><link href="https://adntaha.github.io/blog/whats-ai-safety-even-about/" rel="alternate" type="text/html" title="What’s AI Safety even about?" /><published>2025-08-17T19:43:13-04:00</published><updated>2025-08-17T19:43:13-04:00</updated><id>https://adntaha.github.io/blog/whats-ai-safety-even-about</id><content type="html" xml:base="https://adntaha.github.io/blog/whats-ai-safety-even-about/"><![CDATA[<p>If you pay attention, alongside big AI labs like DeepMind or OpenAI’s <a href="https://www.nature.com/articles/d41586-025-02343-x">releases on the performance of their models on the IMO</a>, you’ll notice that they’ll talk about things like making their AIs ‘safe’, ‘interpretable’ or even ‘trustworthy’ in their blog posts. Or at least that was my case.</p>

<p>And then you tend to wonder, what does this even mean? How do you go about trusting a machine? And with those questions, you’ve entered the field of AI safety.</p>

<p>The issue is that, as AI models get more advanced, people tend to give them more and more responsibility. And it’s a problem because, no matter who you ask, they’ll tell you that we don’t <em>really</em> understand them.</p>

<hr />

<p>Throughout my research, I’ve found that academics and industry refer to individual AIs as “an AI model”, or just “this model”, because they use AI to refer to the whole field. This is a reference to models in the field of statistics, which are basically really fancy equations, and AIs, in a certain sense, can also be seen as really big and fancy equations.</p>

<hr />

<p>Before getting to the first major problem in AI safety, we have to understand what an AI model actually is.</p>

<p>Artificial Intelligence as a word was invented by John McCarthy at what became the first AI conference. It describes, in so little words, a smart computer. Among the different ways we’ve tried to make computers smart is through copying brains. Brains are blobs of inter-connected neurons, and each neuron can receive an electrical signal and emit electrical signals to other neurons. But not all signals are equal: some of them are stronger than others<sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup>. Neurons can also change their connections, so which neurons they’re connected to, and that’s how we learn information.</p>

<p>So in the 1900s, we created a virtual neuron and called it a Perceptron. It’s called that because they were trying to recognize, <em>perceive</em>, what was on a picture, which was basically impossible for computers to do at the time. Each perceptron has two different knobs, but I’ll avoid going into detail because it’s not useful for this exercise.</p>

<p>We can understand modern AI models as a collection of millions, billions or even trillions of Perceptrons. That’s a lot of dials. To train them, we give them huge amounts of information, including huge amounts of books and almost the entire internet for models like ChatGPT or Gemini, and let them turn all the knobs a little bit each time, until we’re satisfied with them. We do this because it’s impossible for a team of humans to turn these nobs themselves. If I was able to tweak 3 knobs per second, it would take me ~2000<sup id="fnref:2" role="doc-noteref"><a href="#fn:2" class="footnote" rel="footnote">2</a></sup> years for a model like GPT-3, just to turn each dial once. And 3 knobs per second is a lot faster than I’d actually be able to achieve. Anyways, as of now, it’s impossible for us to understand what each knob does, since we’re not the ones who tweaked them, and because there’s an incredible amount of them. That’s the first problem of AI safety: we’re unable to <em>interpret</em> what each knob does, how it influences the rest of the knobs—since they’re interconnected!—and so we have no clue how the model works.</p>

<p>The second problem comes from how we train it. Humans are imperfect, and as time passed we got rid of more and more of our stereotypes and prejudice. But since we’re trying to give AI models more and more content to train on, we end up training on some bad stuff which I won’t get into the details of here. You’ve probably seen what’s on the Internet, you know what I’m talking about. Another example is that there are a lot more white male faces on the Internet then black women, or other minorities, which means that face recognition models (much much much smaller than ChatGPT<sup id="fnref:3" role="doc-noteref"><a href="#fn:3" class="footnote" rel="footnote">3</a></sup>) are therefore more accurate on white males than others<sup id="fnref:4" role="doc-noteref"><a href="#fn:4" class="footnote" rel="footnote">4</a></sup>. That’s the second problem of AI safety, which people call algorithmic bias.</p>

<p>The third and fourth problems in AI safety are simpler to understand. Imagine if you had a new job as a cleaner, and your manager taught you to clean by showing you how to mop the floor or wash a shirt, with water, soap, etc. And then a customer comes in and asks you to clean their computer, which you do by washing it under the sink, and you end up breaking it because electronics aren’t supposed to touch water. The thing that the manager did wrong was not telling you how to clean laptops, and the thing you did wrong is that you washed the laptop instead of cleaning it, which would’ve been to open it and blow air inside it to remove the dust. We call the former “out of distribution errors” or “robustness failure”, because you weren’t taught how to clean that specific thing because your manager wasn’t robust with your training. Sorry if the name is weird, academics aren’t exactly known for their naming skills. The latter is called “goal misgeneralisation”, where the customer told you their goal but you didn’t understand it correctly.</p>

<p>That’s basically the gist of it. And to answer the question I asked at the start, AI safety is about making sure that we can trust AI models in sensitive scenarios, like in hospitals for example, while being sure that if something goes wrong, it won’t mess everything up.</p>

<p>If you want to spend more time learning about AI safety, Nicky Case made a really good resource on it—even if it’s not finished yet—that’s available online: <a href="https://aisafety.dance/">AI Safety for Fleshy Humans</a></p>

<div class="footnotes" role="doc-endnotes">
  <ol>
    <li id="fn:1" role="doc-endnote">
      <p>Neurons can emit electrical signals from -40 to -90 millivolts (<a href="https://www.ncbi.nlm.nih.gov/books/NBK11069/">Purves, Augustine, Fitzpatrick, et al. 2001</a>). It’s negative because the inside of the neuron is more negatively charged than the outside. <a href="#fnref:1" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:2" role="doc-endnote">
      <p>GPT-3 has approximately 175 trillion parameters (or nobs), divided by 3 knobs per second, is around 60 trillion seconds, or around 2000 years. <a href="#fnref:2" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:3" role="doc-endnote">
      <p>Just to name an example, FaceNet, which is on the bigger side, has 140 million parameters (<a href="https://learnopencv.com/face-recognition-models/">LearnOpenCV, Durai 2023</a>). That’s 1,000x less than GPT-3! <a href="#fnref:3" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:4" role="doc-endnote">
      <p>According to a study published in 2018 by MIT and Microsoft researchers, the error rate on darker-skinned women is 34.7% as compared to 0.8% on light-skinned men (<a href="https://proceedings.mlr.press/v81/buolamwini18a/buolamwini18a.pdf">Buolamwini and Gebru 2018</a>), which is a huge difference in accuracy. <a href="#fnref:4" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
  </ol>
</div>]]></content><author><name>Aidan Taha</name></author><category term="Decoding AI" /><category term="AI Safety" /><category term="LLMs" /><category term="Medical AI" /><summary type="html"><![CDATA[Where did this field even come from? Why are we worrying about the safety of the things that will replace us? Well, that's not what most AI Safety researchers do, they worry about something else.]]></summary></entry><entry><title type="html">My Path to AI Safety</title><link href="https://adntaha.github.io/blog/my-path-to-ai-safety/" rel="alternate" type="text/html" title="My Path to AI Safety" /><published>2025-06-28T15:52:20-04:00</published><updated>2025-06-28T15:52:20-04:00</updated><id>https://adntaha.github.io/blog/my-path-to-ai-safety</id><content type="html" xml:base="https://adntaha.github.io/blog/my-path-to-ai-safety/"><![CDATA[<p>Hi there. I’m Aidan. If you’ve landed on this blog, then I can safely assume that AI interests you.</p>

<p>In March of 2023, if you don’t know, <a href="https://www.theverge.com/2023/3/8/23629362/meta-ai-language-model-llama-leak-online-misuse">Meta’s Llama model was leaked online</a>. Whether or not it’s a good thing was a hot debate in various internet circles back then, one I didn’t have the knowledge to participate in. All I knew was that, thanks to <a href="https://github.com/ggml-org/llama.cpp">LLaMA.cpp</a>, I had the ability to run it on my computer, and that it would enthusiastically respond to queries asking for the know-how on creating certain molecules and other instructions that would be very bad should they fall into the wrong hands. It was too easy. The risk that these raw, non-safeguarded models can pose is, legitimately, unreal. That being said, Meta’s decisions to release the models–with safeguards in place, this time–feels to me like the only right move<sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup>.</p>

<p>I started looking at AI as a field at the start of a school year, when I had been tasked to find a topic for and get started on a year-long project. That’s when I ran into the <a href="https://www.nature.com/articles/s41597-023-02432-4">FracAtlas dataset</a>, published on Nature<sup id="fnref:2" role="doc-noteref"><a href="#fn:2" class="footnote" rel="footnote">2</a></sup>. This project ended up actually being quite painful, because as it turns out, AI models are spread between a lot of libraries. On top of that, I had to figure out why the models’ predictions were terrible<sup id="fnref:3" role="doc-noteref"><a href="#fn:3" class="footnote" rel="footnote">3</a></sup>, and as any programmer can tell you, debugging complex programs without a debugger is a painful experience.</p>

<p>A lot of articles helped<sup id="fnref:4" role="doc-noteref"><a href="#fn:4" class="footnote" rel="footnote">4</a></sup>, but there’s one thing that I could not figure out: loss functions. Seriously, I didn’t know which one to choose, nevermind the logic behind which models work better than others. Since these models took hours to train, and my <a href="https://research.google.com/colaboratory/faq.html#free-to-use">free Colab instance</a> didn’t like it when I left it to run overnight, I eventually settled with what I had and decided to optimize for other factors<sup id="fnref:5" role="doc-noteref"><a href="#fn:5" class="footnote" rel="footnote">5</a></sup>. The experience, forgive my language, it sucked. It did not feel like the technology of the future that media portrays AI to be.</p>

<p>This frustration with AI black boxes, that the loss was this unknown function that was minimized to try to improve the AI model yet resulted in terrible performance, made me start thinking about bigger questions. If I couldn’t understand my (relatively) simple fracture detection model, how could anyone understand the AI systems making real-world decisions? Or ChatGPT with its multi-hundred-billion parameters? This rabbit hole eventually brought me to 80,000 Hours, a non-profit with the goal of identifying the world’s “most pressing problems” and helping people find careers that work towards solutions to them. So what’s the most pressing problem? At the time of writing, they say it’s power-seeking AI systems. If we use AI to make huge decisions without knowing without any doubt what it’s doing, it could misguide us and lead to catastrophic things. Humanity-ending level things. So how do we avoid this problem? Through AI safety research. Not only does 80,000 Hours classify it as the most pressing problem, but the technical side of this field aligns with my interests.</p>

<p>That’s it. I’ve found what I want to spend my career towards.<sup id="fnref:6" role="doc-noteref"><a href="#fn:6" class="footnote" rel="footnote">6</a></sup> My next goals are an undergrad diploma, master’s and a job at an AI lab if not a doctorate. But that’s a very long term plan. For now, I’ve got my sights set on IJCAI 2025. It’s an incredible opportunity for me to meet people in the field, to hold interesting discussions, and to learn. As a world-class conference, it’s equally good of a place to give people, my peers, others, what I wish I had: a centralized place to learn about AI safety without needing to invest months into it. Explanations of and insights into cutting-edge research, written in a way I can understand. As someone who’s felt the frustration of AI’s black boxes firsthand, I want to help others understand these systems before they become even more powerful.</p>

<div class="footnotes" role="doc-endnotes">
  <ol>
    <li id="fn:1" role="doc-endnote">
      <p>This is the subject of another debate, which is ongoing, even today. As a society, we need to figure this out sooner rather than later because super-intelligent AIs are gonna arrive at some point and we need to have it figured out before then. <a href="#fnref:1" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:2" role="doc-endnote">
      <p>I later found out that Nature is actually a pretty prestigious publication, which only motivated me to get started on this topic even more. <a href="#fnref:2" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:3" role="doc-endnote">
      <p>I mean it. The model accuracy I ended up with was around 55%. <a href="#fnref:3" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:4" role="doc-endnote">
      <p>I don’t quite remember which articles I read, but kudos to all the authors on Medium explaining the various pieces in AI. <a href="#fnref:4" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:5" role="doc-endnote">
      <p>I ended up going with YOLOv8s to allow the model to be ran on a phone. This was enough for the scope of the project. <a href="#fnref:5" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
    <li id="fn:6" role="doc-endnote">
      <p>Statistically speaking, there’s a good chance that I change my mind at some point, but there’s no way to know the future, so I’m going to assume that I’m always going to be interested in AI safety and its problems. <a href="#fnref:6" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
  </ol>
</div>]]></content><author><name>Aidan Taha</name></author><category term="Decoding AI" /><category term="AI Safety" /><category term="Medical AI" /><category term="LLMs" /><category term="Model Failure" /><summary type="html"><![CDATA[The story of a catastrophic failure, a 55% accurate model, and the rabbit hole that led me to the most important problem in tech.]]></summary></entry></feed>