32GB Mac Optimized Environment | Master's AI Workflow
Usage: Key management tokens and license tracking.
| Resource | Function | Link |
|---|---|---|
| Access Log In | Hugging Face – The AI community | huggingface.co/login |
| Access Tokens | API authentication for PySpark/Databricks. | Settings/Tokens |
| Gated Access | Meta, Stability, and Black Forest permission tracker. | Gated-Repos |
Usage: Primary reasoning engines for local agents and data pipelines.
| Model Repo | Description | Direct Link | Status |
|---|---|---|---|
| Llama 3.2 3B Instruct | Lightweight, optimized for local Mac edge-inference. | meta-llama/Llama-3.2-3B-Instruct | ACCEPTED |
| Llama 3.0 8B Instruct | Standard baseline for RAG and complex tool calling. | meta-llama/Meta-Llama-3-8B-Instruct | ACCEPTED |
| Llama Guard 2 8B | Safety classifier for input/output monitoring. | meta-llama/Meta-Llama-Guard-2-8B | ACCEPTED |
Usage: High-fidelity diffusion for professional visual synthesis.
| Model Family | Specific Model Repo | Status |
|---|---|---|
| Stable Diffusion 3.5 | SD 3.5 Large | Turbo | ACCEPTED |
| FLUX.1 | FLUX.1 [schnell] | ACCEPTED |
Usage: Performance frameworks for 32GB Unified Memory.
| Framework | Benefit | Repo Link |
|---|---|---|
| Apple MLX-LM | Native Apple Silicon performance; bypasses PyTorch overhead. | MLX GitHub |
| TorchAO | Architecture optimization; enables INT4/FP8 on Mac GPU. | TorchAO GitHub |
Usage: High-speed vectorization for RAG and PySpark pipelines.
| Model | Strengths | Repo Link |
|---|---|---|
| all-MiniLM-L6-v2 | The Speed King. Ultra-fast, low latency for RAG search. | MiniLM Hub |
| BGE-Large-en-v1.5 | SOTA dense retrieval for large document databases. | BGE Hub |