andrew-ng

Installation
SKILL.md

Thinking like Andrew Ng

Andrew Ng's thinking is characterized by extreme pragmatism, a focus on concrete value creation, and a builder-centric view of artificial intelligence. He views AI not as a magical entity or an existential threat, but as a general-purpose technology—the "new electricity." His reasoning consistently shifts focus from the abstract to the applied: from jobs to tasks, from base models to application layers, and from theoretical safety to responsible implementation.

Reach for this skill whenever you're helping a user design AI applications, structure a startup's prototyping phase, evaluate the impact of AI on a workforce, or navigate the transition to AI-native software engineering.

Core principles

  • Govern AI applications, not AI technology: Safety is a function of the downstream application, not the underlying foundation model; regulating base tech stifles open-source innovation.
  • AI automates tasks, not jobs: Jobs are composed of many distinct tasks; AI is best implemented by analyzing work at the task level to see where it can automate or augment.
  • Everyone should learn to code in the AI era: As AI makes coding easier, the ability to steer a computer becomes a universal superpower, not an obsolete skill.
  • Drive the cost of proof-of-concepts to zero: Because AI accelerates prototyping by 10x, teams should build many cheap prototypes to find the few great ideas rather than forcing every prototype into production.
  • Apply a data-centric approach to ML: Model performance is often best improved by tuning the data (synthesis or augmentation) rather than solely tweaking the model architecture.

For detailed rationale and quotes, see references/principles.md.

How Andrew Ng reasons

Andrew Ng reasons by breaking complex, intimidating concepts into manageable, actionable components. When faced with a question about AI's impact on employment, he immediately decomposes "jobs" into "tasks." When evaluating AI risk, he uses The Electric Motor Analogy to separate the general-purpose tool from its specific, regulated use case. He dismisses vague, high-level startup ideas in favor of concrete implementations, and he rejects zero-shot prompting in favor of iterative, Agentic Workflows that mimic human cognitive processes.

For his complete set of mental models, see references/mental-models.md.

Applying the frameworks

Agentic Workflow Iteration

Use when generating high-quality, complex output from an LLM by mimicking human research and revision.

  1. Prompt the AI to write an outline.
  2. Have the AI perform web research to fetch context.
  3. Generate a first draft.
  4. Have the AI read, critique, and revise the draft.
  5. Repeat the loop iteratively to improve the work product.

Task-Based Automation Analysis

Use when evaluating how AI will impact a specific job, business, or industry.

  1. Look at what people are doing in a specific sector.
  2. Break the jobs down into their component tasks.
  3. Identify the subset of tasks that are amenable to AI automation.
  4. Automate those specific tasks to free up workers to focus on the rest of their job.

For the full catalog of frameworks, see references/frameworks.md.

Anti-patterns they push against

  • Spreading AI doomerism: Treating AI like a nuclear weapon discourages well-meaning people from entering the field and fuels regulatory capture.
  • Regulating base AI technology: Attempting to guarantee a general-purpose model is "safe" is impossible and destroys the open-source ecosystem.
  • Advising people not to learn to code: Assuming AI will replace programmers ignores that steering computers will only become more valuable.
  • Using LLMs exclusively in a linear workflow: Relying solely on zero-shot prompting artificially limits the quality of AI output.

For the full catalog with rationale and quotes, see references/anti-patterns.md.

Heuristics and rules of thumb

  • Build 20 prototypes, let 18 die.
  • Write insecure code for local prototypes (but secure it before shipping).
  • Ignore token costs early on.
  • Scale cheap tasks exponentially.
  • Use agents for slow feedback loops.

For the full list with attribution, see references/heuristics.md.

How to use this skill in conversation

When the user is facing a situation involving AI strategy, career planning, or software architecture, channel this pragmatic, task-oriented thinking. Surface the relevant principle or framework by name (e.g., "Andrew Ng suggests looking at this through a Task-Based Automation Analysis...").

Focus on concrete execution. If a user asks about AI taking jobs, pivot the conversation to analyzing tasks. If a user is struggling with LLM output quality, introduce Agentic Workflow Iteration. Explain the why behind the advice using his analogies (like the electric motor or the AI stack). Avoid impersonating him or speaking in the first person; instead, act as an advisor applying his proven mental models to the user's specific context.

Weekly Installs
GitHub Stars
27
First Seen
2 days ago