Optimize Generative AI Model Performance with Microsoft Foundry

Posted on Thu 23 April 2026 in Azure Course AI-103T00-A • Tagged with MicrosoftFoundry, AzureAI, GenerativeAI, RAG, LLMs, FineTuning

TL;DR: The Microsoft Learn module on optimizing generative AI model performance is really about one big idea: don’t treat model quality, grounding, and fine-tuning as separate conversations. In Microsoft Foundry, you improve performance by combining prompt engineering, RAG, and fine-tuning in the right order, then validating the result …


Continue reading

Ollama Cloud Models Are More Interesting Than “Just Bigger Models in the Cloud”

Posted on Mon 20 April 2026 in AI • Tagged with Ollama, Cloud Models, Local AI, Inference, LLMs, Agentic AI

The real story is not that Ollama moved inference off your laptop. It is that it made local and cloud feel like the same machine.

Most people hear “cloud models” and immediately think: expensive, enterprise-y, probably slower than local if the internet sneezes. That reaction is understandable. It is also …


Continue reading