AI EducademyAIEducademy
🌳

مسار تعلّم الذكاء الاصطناعي

🌱
AI Seeds

Start from zero

🌿
AI Sprouts

Build foundations

🌳
AI Branches

Apply in practice

🏕️
AI Canopy

Go deep

🌲
AI Forest

Master AI

🔨

مسار هندسة البرمجيات

✏️
AI Sketch

Start from zero

🪨
AI Chisel

Build foundations

⚒️
AI Craft

Apply in practice

💎
AI Polish

Go deep

🏆
AI Masterpiece

Master AI

عرض كل البرامج→

المختبر

تم تحميل 7 تجارب
🧠ملعب الشبكة العصبية🤖ذكاء اصطناعي أم إنسان؟💬مختبر التوجيهات🎨مولّد الصور😊محلل المشاعر💡باني روبوت الدردشة⚖️محاكي الأخلاقيات
دخول المختبر→
📝

المدونة

أحدث المقالات في الذكاء الاصطناعي والتعليم والتكنولوجيا

اقرأ المدونة→
الأسئلة الشائعة
🎯
رسالتنا

جعل تعليم الذكاء الاصطناعي متاحاً للجميع في كل مكان

💜
قيمنا

مفتوح المصدر، متعدد اللغات، وقائم على المجتمع

⭐
مفتوح المصدر

مبني علناً على GitHub

تعرّف على المنشئ→عرض على GitHub
ابدأ الآن
AI EducademyAIEducademy

رخصة MIT. مفتوح المصدر

تعلّم

  • البرامج الأكاديمية
  • الدروس
  • المختبر

المجتمع

  • GitHub
  • المساهمة
  • قواعد السلوك
  • عن المنصة
  • الأسئلة الشائعة

الدعم

  • اشترِ لي قهوة ☕
البرامج الأكاديمية للذكاء الاصطناعي والهندسة›🏆 AI Masterpiece›الدروس›قراءة الأوراق البحثية
📄
AI Masterpiece • متقدم⏱️ 20 دقيقة للقراءة

قراءة الأوراق البحثية

Reading Research Papers - How to Understand and Implement AI Papers

The gap between "I use AI" and "I build AI" is bridged by reading research papers. Every breakthrough - transformers, diffusion models, RLHF - started as a paper. Engineers who read papers don't just follow trends; they anticipate them.

🤔 Why Read Papers?

  • Stay current - blog posts lag behind papers by months
  • Understand fundamentals - tutorials simplify; papers explain why
  • Debug better - when your model fails, the paper reveals edge cases
  • Career leverage - paper-literate engineers are rare and valued
The three-pass paper reading approach showing skim, understand, and critique stages
The three-pass method transforms a dense 12-page paper into structured understanding.

📐 Anatomy of an ML Paper

Every ML paper follows a predictable structure. Knowing the blueprint accelerates reading:

| Section | Purpose | Time to Spend | |---------|---------|--------------| | Abstract | 200-word summary of the entire contribution | 2 minutes | | Introduction | Problem motivation, why existing solutions fail | 5 minutes | | Related Work | What came before, how this paper differs | Skim on first pass | | Method | The core contribution - architecture, algorithm, maths | 60% of your time | | Experiments | Proof that it works - datasets, baselines, ablations | 20% of your time | | Conclusion | Summary and future directions | 2 minutes |

\ud83e\udd2f
The original "Attention Is All You Need" paper (2017) has been cited over 130,000 times, making it one of the most influential computer science papers ever written. Its title has spawned countless parodies in subsequent paper titles.

📖 The Three-Pass Approach

Pass 1: Skim (15 minutes)

Read the abstract, introduction, section headings, figures, and conclusion. After this pass, you should be able to answer:

  • What problem does this paper solve?
  • What is the key idea in one sentence?
  • Is this paper relevant to my work?

If the answer to the third question is no, stop here. Not every paper deserves a deep read.

Pass 2: Understand (1–2 hours)

Read the full paper, skipping dense proofs on first encounter. Focus on:

  • Figures and tables - authors distil their best ideas into visuals
  • Method section - trace the data flow from input to output
  • Ablation studies - these reveal which components actually matter

Mark sections you don't understand. Move on and return to them after reading the experiments - results often clarify the method.

Pass 3: Critique (1–2 hours)

Now read as a reviewer. Ask:

  • What assumptions does this paper make? Are they reasonable?
  • What's missing from the experiments? Which baselines are absent?
  • How would this approach fail? What are its limitations?
  • Can I reproduce these results with the information provided?
\ud83e\udde0فحص سريع

During Pass 2 of reading a paper, you encounter a mathematical derivation you don't understand. What's the BEST next step?

📚 Essential Papers Every AI Engineer Should Read

These papers form the foundation of modern AI - reading them is non-negotiable:

Architectures:

  • Attention Is All You Need (Vaswani et al., 2017) - the transformer
  • Deep Residual Learning for Image Recognition (He et al., 2015) - ResNet and skip connections

Language Models:

  • BERT: Pre-training of Deep Bidirectional Transformers (Devlin et al., 2018) - bidirectional pre-training
  • Language Models are Few-Shot Learners (Brown et al., 2020) - GPT-3 and in-context learning

Training Techniques:

  • Adam: A Method for Stochastic Optimisation (Kingma & Ba, 2014) - the default optimiser
  • Dropout: A Simple Way to Prevent Neural Networks from Overfitting (Srivastava et al., 2014)
\ud83e\udd14
Think about it:If you could only read three papers from the list above, which three would give you the broadest understanding of modern AI? Why those three?

🔢 Reading Maths Notation

Mathematical notation is a language. Here's your phrasebook for ML papers:

θ (theta)      - model parameters (weights)
∇ (nabla)      - gradient operator
𝔼 (E)          - expected value (average over a distribution)
argmax          - the input that maximises a function
∑ (sigma)      - summation
∏ (pi)         - product
‖x‖            - norm (magnitude) of vector x
softmax(z_i)   - e^(z_i) / Σ e^(z_j) - converts logits to probabilities

Pro tip: when you see a complex equation, substitute actual numbers. If the paper says L = -Σ y_i log(ŷ_i), plug in y = [1, 0] and ŷ = [0.9, 0.1] and compute by hand. Abstraction becomes concrete instantly.

💻 From Paper to Code

The ultimate test of understanding: implement it. A practical workflow:

# Step 1: Implement the core mechanism in isolation
import torch
import torch.nn as nn

class SelfAttention(nn.Module):
    """Scaled dot-product attention from 'Attention Is All You Need'"""
    def __init__(self, d_model: int, n_heads: int):
        super().__init__()
        self.d_k = d_model // n_heads
        self.n_heads = n_heads
        self.W_q = nn.Linear(d_model, d_model)
        self.W_k = nn.Linear(d_model, d_model)
        self.W_v = nn.Linear(d_model, d_model)

    def forward(self, x):
        B, T, C = x.shape
        q = self.W_q(x).view(B, T, self.n_heads, self.d_k).transpose(1, 2)
        k = self.W_k(x).view(B, T, self.n_heads, self.d_k).transpose(1, 2)
        v = self.W_v(x).view(B, T, self.n_heads, self.d_k).transpose(1, 2)

        # Scaled dot-product attention (Equation 1 in the paper)
        attn = (q @ k.transpose(-2, -1)) / (self.d_k ** 0.5)
        attn = torch.softmax(attn, dim=-1)
        return (attn @ v).transpose(1, 2).contiguous().view(B, T, C)

Verify against the paper: check tensor shapes at each step. If the paper says output is (batch, seq_len, d_model), assert that.

\ud83e\udde0فحص سريع

What is the BEST way to verify your paper implementation is correct?

🔍 Finding Papers

| Source | Best For | |--------|----------| | arXiv (arxiv.org) | Latest preprints, fastest access | | Semantic Scholar | Citation graphs, finding related work | | Papers With Code | Papers linked to implementations and benchmarks | | Connected Papers | Visual exploration of paper relationships | | Twitter/X | Real-time discussion of new papers |

\ud83e\udd2f
arXiv receives over 2,000 new submissions per day across all fields. The cs.LG (Machine Learning) and cs.CL (Computation and Language) categories alone account for hundreds of daily papers.
\ud83e\udde0فحص سريع

Which resource is MOST useful when you want to find an existing code implementation of a paper's method?

\ud83e\udd14
Think about it:Start a paper reading habit: one paper per week, summarised in three sentences (problem, method, result). After 12 weeks, you'll have read more papers than most working ML engineers read in a year.

🎯 Key Takeaways

  • Use the three-pass method: skim, understand, critique
  • Focus 60% of your reading time on the method section
  • Substitute real numbers into equations to build intuition
  • Implement the core mechanism to truly understand it
  • One paper per week compounds into expertise

📚 Further Reading

  • How to Read a Paper by S. Keshav - The original three-pass method paper, widely cited in academia
  • Papers With Code - Browse papers with linked implementations and leaderboards
  • The Illustrated Transformer by Jay Alammar - Visual walkthrough of the transformer architecture
الدرس 8 من 100٪ مكتمل
←دليل مسابقات Kaggle
محاكاة مقابلة تصميم الأنظمة→