[tutorial] Introduction to sigstore's model signing tool

Note: This page is an AI-generated (gpt-5-mini-2025-08-07) translation from Traditional Chinese and may contain minor inaccuracies.

📌 Introduction

Introducing Sigstore’s model signing tool and describing how its three-layer structure (Sigstore Bundle, DSSE Envelope, In‑Toto Statement) provides integrity, authenticity, and transparency for machine learning model signing and verification. Also demonstrates simple signing and verification commands.

Read more

[note] Introduction to SBOM, AIBOM and Their Tools

Note: This page is an AI-generated (gpt-5-mini-2025-08-07) translation from Traditional Chinese and may contain minor inaccuracies.

📌 Introduction

This article introduces the Software Bill of Materials (SBOM) and its application in the AI domain, AIBOM. It briefly presents three common SBOM standard formats: SPDX, CycloneDX, and SWID. While introducing AIBOM, it explains the differences between AIBOM and Model Cards, and presents the open-source tool aetheris-ai/aibom-generator, which can extract model information from Hugging Face and perform scoring.

Read more

[note] Introduction to automated AI security auditing framework petri

Note: This page is an AI-generated (gpt-5-mini-2025-08-07) translation from Traditional Chinese and may contain minor inaccuracies.

📌 Introduction

Petri is a red-team tool for AI safety testing that simulates realistic interactive scenarios to detect potential model risks. Through collaboration between the Auditor, Target, and Judge, it performs various tasks such as general audits, multi-model comparisons, and whistleblowing tests to check whether models leak information, exhibit bias, or show other issues, improving AI safety and reliability in complex scenarios.

Read more