Home
Unchained
Product Blog

Securing the ML supply chain with new Chainguard AI Images

Dan Lorenc, CEO

While popular AI languages like TensorFlow, PyTorch and Kubeflow are great for data science use cases, they can pose challenges when being deployed into security critical environments because of their large size, attack surface, and package management issues.

Today, we are announcing a new Chainguard Images AI bundle that will include a comprehensive collection of images for stages in the AI workload lifecycle, from development images to workflow management tools, to vector databases for production storage. As with all of our Images, they include critical software supply chain security features like software signatures, SBOMs and CVE remediation. At Chainguard, we believe that developers shouldn’t have to choose between easy-to-use and secure, and our new collection of AI Images is another step in that direction. Today’s new Chainguard Images AI bundle includes:

  • Python, Conda, OpenAI and Jupyter notebook images for developing models and using the OpenAI API,


  • Kubeflow images for deploying production ML pipelines to Kubernetes-based platforms,


  • Milvus and Weaviate vector database images for data storage.

Take preventative steps to tackle AI supply chain threats

More than 53 percent of organizations are aggressively pursuing large language applications, and Docker recently reported an astounding 100 million pull requests for AI/ML Images on Docker Hub. As the recent PoisonGPT attack technique demonstrated just how ripe LLMs are for security exploits, raising concerns about what the inevitable large-scale exploit could look like when it hits the LLM software supply chain. AntidoteGPT was introduced soon after the PoisonGPT article findings to prevent this type of attack vector by applying techniques from Chainguard Images and the Sigstore project.


Large Language Models (LLMs) provide a powerful primitive that can be used to automate many tasks in our daily lives with shocking accuracy and efficiency. Data science workflows are at a similar evolutionary phase as software development was 20 years ago, where no automated testing, code review, or even version control were the norm. Our Chainguard Labs team recently published lessons learned from securing container operating systems and how these principles apply to AI/ML infrastructure. As AI/ML workloads begin to move past chatbots and into more sensitive workloads, the security of the infrastructure they run on will begin to matter more and more. That’s why it was important to us to create a collection of hardened Chainguard Images dedicated to running these workloads.

How it works

As always, our Chainguard Images are based on Wolfi, our community un-distro, which means they’re hardened by default, a fraction of the size, and aim to meet our standard zero-known CVE SLA through daily updates and patching.


Comparisons for Chainguard AI Images as of August 1, 2023.

You can try Chainguard Images today to see for yourself how we're working to improve the container image landscape with a secure-by-default design. Our free and public Images are available on the :latest and :latest-dev: versions only. If you're interested in other versions in today’s AI Images bundle, reach out to our team for more information. Our Chainguard Images inventory is always expanding. If you need something you don’t see listed in our catalog, let us know.

Interested in seeing how we approach building our Images and what makes them more secure than the alternative options? Watch our CEO Dan Lorenc build Chainguard Images here:



Share

Ready to Lock Down Your Supply Chain?

Talk to our customer obsessed, community-driven team.

Get Started