LLM
loading
·
loading
·
Base Mac Mini M4, The alternatives for Low-End NVIDIA Hardware for Inference
·621 words·3 mins
How the Mac Mini M4 has enabled affordable Local LLM inference and making VRAM-starved NVIDIA cards obsolete for low power and low cost
The Strawberry Challenge, When LLMs Need Tools to Count
·170 words·1 min
Around October 2024, The infamous “How many R’s are in strawberry?” question has become a fascinating litmus test for Large Language Models, exposing a fundamental limitation in how these systems process text.
AI in Finance Workshop with a live demo
·295 words·2 mins
Artificial Intelligence Applications in Financial Services with a demo on Local LLM Invoice Processing with Automated Billing Code Assignment
Concurrency in LLMs - Why It Matters More Than size of LLM
·321 words·2 mins
Understanding why handling multiple requests beats raw token speed for my local LLM deployments.