Generate viral LinkedIn posts in your style for free.

Generate LinkedIn posts
Paolo Perrone

Paolo Perrone

These are the best posts from Paolo Perrone.

8 viral posts with 3,955 likes, 693 comments, and 360 shares.
5 image posts, 0 carousel posts, 1 video posts, 1 text posts.

๐Ÿ‘‰ Go deeper on Paolo Perrone's LinkedIn with the ContentIn Chrome extension ๐Ÿ‘ˆ

Best Posts by Paolo Perrone on LinkedIn

OpenAI rejects 99% of AI engineers.
The 1% who get in? They all watched the same YouTube playlist.

It's Karpathy's Neural Networks: Zero to Hero.

Most courses teach you to call APIs. Karpathy teaches you to be a builder:
โ†’ Implements backprop from scratch (finally understood it)
โ†’ Builds GPT from bare NumPy (no magic imports)
โ†’ Shows the actual math without the fluff
โ†’ Live codes everything (mistakes included)

The difference is brutal:

Course graduates: "How do I import torch.nn?"
Karpathy students: "Here's how attention actually works."

Guess who gets hired?

Start with micrograd. Thank me at your OpenAI interview.
Full playlist in the comments ๐Ÿ‘‡
Post image by Paolo Perrone
MIT Press will charge $80 for this textbook in 2026.

You're reading it for free today.

"Machine Learning Systems" โ€” The missing education between algorithms and production.

What's inside (save this):

Part 1: Foundations (4 chapters)
โ†’ Data pipelines that don't break at scale
โ†’ Training systems beyond Jupyter notebooks
โ†’ Benchmarking that predicts real performance
Skip this โ†’ Your models die in production

Part 2: Deployment (4 chapters)
โ†’ Edge AI on $10 hardware (TinyML)
โ†’ Cloud patterns that actually scale
โ†’ MLOps beyond "it works locally"
Skip this โ†’ Inference costs eat your runway

Part 3: Advanced Systems (4 chapters)
โ†’ Distributed training that works
โ†’ Hardware acceleration (GPUs โ†’ TPUs โ†’ custom)
โ†’ Model compression without accuracy loss
Skip this โ†’ You're stuck at prototype scale

Part 4: Responsible AI (4 chapters)
โ†’ Privacy systems, not just policies
โ†’ Carbon footprint measurement
โ†’ Security against adversarial attacks
Skip this โ†’ Your AI becomes a liability

Bonus: TinyTorch framework included.
Build neural networks from scratch. No black boxes.

10,300 engineers starred it.
Prof. Vijay Janapa Reddi updates it weekly.
MIT Press publishes it 2026.

You're reading it today.

๐Ÿ“– Read: mlsysbook.ai
๐Ÿ“„ PDF: mlsysbook.ai/pdf
๐Ÿ”ฌ GitHub: https://lnkd.in/ef_SKs6G

Time to complete: 8 weeks, 2 hours/day.

๐Ÿ’พ Save this before it costs $80
โ™ป๏ธ Repost before someone pays MIT Press for what's free today.
Post image by Paolo Perrone
David Kimai just open-sourced the Context Engineering handbook.

7.1K stars in 2 weeks.

Here's what's actually inside:

The Core Idea:
Prompt engineering = what you type
Context engineering = everything else the model sees

Most people optimize the prompt.
Smart people optimize the context.

What You Get:

๐Ÿ“ 00_foundations/
Basic theory in plain English.
Why context beats prompts every time.
Token budgets that don't blow up.

๐Ÿ“ 10_guides_zero_to_hero/
Start here. Literally.
Minimal examples that actually run.
No 500-line boilerplate.

๐Ÿ“ 20_templates/
Copy-paste context patterns.
Memory systems. Tool integration. Control flow.
YAML configs you can steal.

๐Ÿ“ 30_examples/
Full implementations:
- Chatbots with real memory
- Agents that don't hallucinate
- RAG that actually retrieves

๐Ÿ“ 40_reference/
Deep dives on:
- Why attention windows matter
- Context pruning strategies
- Evaluation metrics that work

The Practical Stuff:

Every concept has:
โœ“ Runnable Python code
โœ“ ASCII diagrams
โœ“ Before/after metrics
โœ“ "Why this matters" sections

No slides. No theory dumps.
Just: problem โ†’ solution โ†’ code.

My Favorite Parts:

The biological metaphor:
atoms โ†’ molecules โ†’ cells โ†’ organs
(single prompts โ†’ few-shot โ†’ memory โ†’ multi-agent)

The token calculator:
Shows exactly why your context explodes.
And how to fix it.

The "cognitive tools" templates:
Prompt programs that make models think step-by-step.
No training required.

Who This Is For:

- Engineers tired of prompt tweaking
- Anyone hitting token limits
- Teams building production agents
- People who want stuff that works

The repo assumes you know Python.
Everything else is explained.

Start with 10_guides/01_min_prompt.py
Run it. Break it. Understand it.

Then steal whatever you need.

https://lnkd.in/gpWCq58e

What part of context engineering confuses you most?

โ™ป๏ธ Repost to help someone graduate from prompt engineering
Post image by Paolo Perrone
The only Agentic AI roadmap you need for 2026.

No fluff. Just what works. With actual links.

Phase 1๏ธโƒฃ: Foundations (2 weeks)
โ†’ Math: 3Blue1Brown Linear Algebra: https://lnkd.in/ewiPRVuG
โ†’ Python basics: https://lnkd.in/eDSYRAkg
โ†’ ML fundamentals: https://lnkd.in/eYZfefYP
Skip this โ†’ You'll be learning theory forever

Phase 2๏ธโƒฃ: Build Your First Agent (2 weeks)
โ†’ ReAct pattern tutorial: react-lm.github.io
โ†’ LangChain quickstart: https://lnkd.in/eZCZHnv7
โ†’ Build memory + tools: https://lnkd.in/e53rpuev
Project: Agent that searches web + executes code
Skip this โ†’ You'll never understand why agents fail

Phase 3๏ธโƒฃ: Advanced Architectures (2 weeks)
โ†’ Multi-agent systems: https://lnkd.in/ganTtyg7
โ†’ AutoGPT architecture: https://lnkd.in/gQBfXtnf
โ†’ RLHF fundamentals: huggingface.co/blog/rlhf
Project: Agent that improves its own prompts
Skip this โ†’ Your agents stay shallow forever

Phase 4๏ธโƒฃ: Production Systems (2 weeks)
โ†’ FastAPI deployment: fastapi.tiangolo.com
โ†’ Docker + agents: https://lnkd.in/eb4tmubv
โ†’ LangSmith monitoring: smith.langchain.com
Reality check: 90% of "AI agents" die here
Skip this โ†’ Your demo stays a demo

Phase 5๏ธโƒฃ: Pick ONE Specialization
๐Ÿค– Robotics: https://lnkd.in/e5sA7XnS
๐Ÿ’ผ Business: docs.crewai.com
๐Ÿ”ฌ Research: paperswithcode.com

Best resources that actually deliver:
๐Ÿ“š Theory: https://lnkd.in/envUC_aC
๐Ÿ›  Practice: python.langchain.com/docs
๐Ÿ”ฅ Deep understanding: mlsysbook.ai/tinytorch
๐Ÿš€ Deploy: railway.app or vercel.com

Time: 8 weeks, 3 hours/day.
Cost: $0 (all resources free).

The difference:
Others: 50 tutorials โ†’ Maybe build something
You: 5 working agents โ†’ Understand everything

๐Ÿ’พ Save this before 2026 hits and you're still "planning to learn agents"
โ™ป๏ธ Repost if someone in your network has been "about to start" for 6 months
Elon Musk tweeted about Cartesia voice AI, so I tested it against 11Labs and OpenAI.

Sent it to 50 people asking which sounded most human.
The results shocked me.

Sonic-3 just changed the game.

3-5x faster than OpenAI.
More accurate than ElevenLabs.
$100M in funding because it actually works.

But here's the insane part:

IT LAUGHS.

Like, actually laughs. Not robot "ha ha ha."
Real, contextual, human laughter.

And it speaks 42 languages. Including 9 Indian languages.
Hindi, Tamil, Telugu - all native fluency.
One voice. Every language. Zero accent bleeding.

The features that blew my mind:

๐ŸŽฏ SPEED CONTROL THAT WORKS
"Say that slower" โ†’ It actually slows down
"Say that faster" โ†’ Speeds up mid-sentence
No other TTS can do this. I tried them all.

๐Ÿ”ฅ VOICE CLONING IN 3 SECONDS
Not 30 minutes of recording. Not 3 minutes.
3. Fucking. Seconds.

๐ŸŒ HANDLES THE IMPOSSIBLE
Email addresses: john.smith_92@outlook.com โœ“
Heteronyms: "Present the present" โœ“
Indian names: "Venkatasubramanian" โœ“
Your current TTS would stroke out.

This isn't incremental improvement. This is voice AI that ships and converts.
โ™ป๏ธ Repost if you're done with voice AI that sounds dead inside
A Fortune 500 CEO typed one sentence.

"Migrate all cloud environments from AWS to Azure.
End to end. Secure Compliant. Human review on critical steps."

That's the input.

Here's what happened:

โ†’ Full project scoped in minutes
โ†’ Agents assigned by role (infrastructure, security, compliance)
โ†’ Tasks executed in parallel, maintaining state
โ†’ Human checkpoints for critical decisions
โ†’ Migration completed overnight

No sprint planning.
No hiring.
No "we'll get to it in Q3."

This is engineering velocity.

Same team. 10x output.
Scale capacity without scaling headcount.

The platform that enables it?

Kubiya

Not an AI that autocompletes your code.
An AI org that ships your roadmap.

The question isn't whether agentic engineering teams are coming.

It's whether you'll deploy and manage them โ€” or they'll replace you.

๐Ÿ‘พ https://www.kubiya.ai/
๐Ÿ“„ https://lnkd.in/e58M_PPw

๐Ÿ’พ Save this for the next "we need to prioritize" meeting that changes nothing
โ™ป๏ธ Repost if your backlog has a backlog
30B parameters running on 24GB. Not a typo.

NVIDIA AI dropped a banger MoE model.

Nemotron 3 Nano.

Runs on 24GB. Only 3.6B active during inference. 1M context window.

I ran it on my DGX Spark. Here's the verdict:

๐—ฆ๐—ฒ๐˜๐˜‚๐—ฝ: โ†’ Clone llama.cpp โ†’ Build with CUDA โ†’ Pull the GGUF from Hugging Face โ†’ Running in 20 minutes

๐—ช๐—ต๐—ฎ๐˜ ๐˜€๐˜‚๐—ฟ๐—ฝ๐—ฟ๐—ถ๐˜€๐—ฒ๐—ฑ ๐—บ๐—ฒ:
โ†’ 3.6B active params competing with models 3x the size
โ†’ Built-in reasoning with tokens โ€“ no prompt hacks
โ†’ Native tool calling. No gymnastics.

๐˜ƒ๐˜€. ๐—ผ๐˜๐—ต๐—ฒ๐—ฟ ๐—น๐—ผ๐—ฐ๐—ฎ๐—น ๐—บ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€:
โ†’ Llama 3: More VRAM, no native reasoning
โ†’ Qwen: Close, but loses on coding benchmarks
โ†’ Mistral Large 3: Similar speed, but 1/4 the context

๐—ง๐—ต๐—ฒ ๐—ด๐—ผ๐˜๐—ฐ๐—ต๐—ฎ:
Watch your context size. I started at 1M and hit OOM. Dial it back to 32K-64K unless you've got headroom.

๐—ค๐˜‚๐—ถ๐—ฐ๐—ธ ๐˜€๐˜๐—ฎ๐—ฟ๐˜:
./llama.cpp/llama-cli \\
-hf unsloth/Nemotron-3-Nano-30B-A3B-GGUF:UD-Q4_K_XL \\
--jinja --ctx-size 32768 \\
--temp 0.6 --top-p 0.95

This is the best "run it on your own hardware" model I've used.

๐Ÿ’พ Save for your next "which local model?" decision
โ™ป๏ธ Repost if your AI stack is going fully open source in 2026
Post image by Paolo Perrone
AI agents wonโ€™t fix your data science workflow. Unless you give them the right environment.

This is how we deliver Data Science projects in 2025 in seven steps:

Start with a prompt, not a blank notebook.
Let the Zerve agent generate the initial workflow and code.
Review the plan and data previews in the full IDE.
Iterate with natural language; the agent stays context-aware.
Scale experiments from one run to thousands instantly.
Trust that every result is tracked and reproducible.
Deploy your work as a workflow, API, or app.

Get Zerve here: https://bit.ly/41i1A61

AI agents won't replace data scientists.
But a data scientist using an agentic environment like Zerve will.
Post image by Paolo Perrone

Related Influencers