llm-engineering-skills

agent
Security Audit
Pass
Health Pass
  • License — License: MIT
  • Description — Repository has a description
  • Active repo — Last push 0 days ago
  • Community trust — 24 GitHub stars
Code Pass
  • Code scan — Scanned 1 files during light audit, no dangerous patterns found
Permissions Pass
  • Permissions — No dangerous permissions requested
Purpose
This project provides a curated collection of specialized engineering skills and knowledge bases designed to enhance AI coding agents (like Claude Code and OpenAI Codex) for machine learning and LLM development tasks.

Security Assessment
Overall Risk: Low. The automated code scan checked the repository files and found no dangerous patterns, hardcoded secrets, or requests for dangerous system permissions. Because the tool acts primarily as a knowledge base and prompt guide for your local AI agent, it does not inherently execute shell commands, make independent network requests, or access sensitive data itself. Users should always exercise standard caution when installing third-party packages via external community CLIs (e.g., `npx`), but the underlying repository content is clean and safe.

Quality Assessment
The project demonstrates strong health and active maintenance. It was updated very recently (as indicated by the 0 days since the last push) and has garnered a respectable 24 GitHub stars, suggesting a growing and interested user base. Furthermore, it is fully transparent and legally clear to use, distributed under the standard and permissive MIT license.

Verdict
Safe to use.
SUMMARY

LLM Engineering Claude Skills

README.md

LLM Engineering Skills

License: MIT
Claude
GitHub Stars
Last Commit
PRs Welcome

Turn AI agents into capable LLM engineers.

This repository is a curated collection of practical engineering skills designed to make AI agents like Claude Code and OpenAI Codex dramatically more effective when working on AI and machine learning projects. Instead of generic assistance, these skills give agents concrete knowledge, workflows, and patterns used by real LLM engineers.

The goal is to enable AI agents to reason, plan, and execute AI engineering tasks with confidence and precision.

Getting Started

Claude Code

Prerequisites

  • Claude Code CLI (version 1.0.33 or later)

Add the plugin marketplace and install:

# Add the marketplace
/plugin marketplace add itsmostafa/llm-engineering-skills

# Install the plugin
/plugin install llm-engineering-skills@itsmostafa-llm-engineering-skills

Or install directly using the community CLI:

npx claude-plugins install @itsmostafa/llm-engineering-skills

Codex

Prerequisites

  • Codex CLI

Install a specific skill using the skill installer:

$skill-installer install https://github.com/itsmostafa/llm-engineering-skills/tree/main/skills/<skill-name>

For example, to install the rlhf skill:

$skill-installer install https://github.com/itsmostafa/llm-engineering-skills/tree/main/skills/rlhf

Usage

Once installed, AI agents will automatically use these skills when you work on relevant tasks. Here are some examples:

Fine-tune a 70B model on a single GPU

"Help me QLoRA fine-tune Gemme 3 24B on my custom dataset using a 48GB A6000"

The agent knows NF4 quantization, double quantization for memory savings, paged optimizers to handle memory spikes, and the exact BitsAndBytesConfig settings to make it work.

Build an AI agent with tool use

"Create a ReAct agent that can search the web, read files, and execute code to answer research questions"

The agent understands orchestrator-worker patterns, human-in-the-loop checkpoints, and how to design tools that return self-contained, LLM-friendly outputs.

Align a model with human preferences

"Implement DPO training to align my instruction-tuned model using preference data"

The agent knows the Bradley-Terry model, when to use DPO vs PPO, KL regularization to prevent reward hacking, and how to structure preference datasets.

Run models locally on Apple Silicon

"Convert Mistral 7B to MLX format and fine-tune it on my M4 Max with LoRA"

The agent handles model conversion, 4-bit quantization for MLX, and memory-efficient LoRA training optimized for unified memory.

Manage context in long-running agents

"My agent loses track of earlier decisions after 50+ turns. How do I fix this?"

The agent implements hybrid context strategies—trimming old messages while maintaining structured summaries—and designs tools for just-in-time context loading.

Optimize training performance

"My PyTorch training is slow. Help me profile it and add torch.compile with the right backend"

The agent knows torch.profiler, when to use inductor vs cudagraphs, gradient checkpointing trade-offs, and how to identify bottlenecks in data loading.

Skills Included

Skill Description
agents Patterns and architectures for building AI agents and workflows. Tool use, multi-step reasoning, and orchestration of LLM-driven tasks.
context-engineering Managing LLM context windows in AI agents. Long conversations, multi-step tasks, and maintaining coherence across extended interactions.
lora Parameter-efficient fine-tuning with Low-Rank Adaptation. Train models with ~0.1% of original parameters using adapter merging.
mlx Running and fine-tuning LLMs on Apple Silicon with MLX. Model conversion, quantization, LoRA fine-tuning, and local model serving.
prompt-engineering Crafting effective prompts for LLMs. Designing prompts, improving output quality, and structuring complex instructions.
pytorch Building and training neural networks with PyTorch. Training loops, data pipelines, torch.compile optimization, and distributed training.
qlora Memory-efficient fine-tuning with 4-bit quantization and LoRA adapters. Fine-tune large models (7B+) on consumer GPUs with limited VRAM.
rlhf Reinforcement Learning from Human Feedback for aligning language models. Reward modeling, policy optimization, and DPO.
transformers Loading and using pretrained models with Hugging Face Transformers. Pipeline API, Trainer fine-tuning, and multimodal tasks.

Contributing

Contributions are welcome.

If you have practical LLM engineering knowledge, workflows, or patterns that would help AI agents perform better on real projects, feel free to open a pull request.

Please update the REFERENCES.md file to include any external references you've used.

License

MIT

Reviews (0)

No results found