
The MCP protocol is almost one year old and during that time, developers have built thousands of new MCP servers.Thinking back to MCP demos from six months ago, most developers were using one or two local MCP servers, each contributing just a handful of tools.Six months later and we have access to thousands of tools, and a new set of issues.
Which MCP servers do we trust? How do we avoid filling our context with tool definitions that we won’t end up needing?
One of the most exciting advances in modern AI is multimodal
support, the ability for models to understand and generate multiple
types of input, such as text, images, or audio. With
multimodal models, you’re no longer limited to typing prompts;you
can show an image or play a sound, and the model can understand
it.This opens a world of new possibilities for developers building
intelligent, local AI experiences.
My name is Mike Coleman, a staff solution architect at Docker.This
year I decided to turn a Home Depot animatronic skeleton into an
AI-powered, live, interactive Halloween chatbot.The
result:kids walk up to Mr.Bones, a spooky skeleton in my yard, ask
it questions, and it answers back — in full pirate voice — with
actual conversational responses, thanks to a local LLM powered by
