Pull Request Review (WIP): ietf-scitt#18
Collection of metric data into shared database (crowdsourcable OpenSSF Metrics). There are many repos to search, we want to enable self reporting and granularity as applicable to ad-hoc formed policy as desired by end-user.
- Related: https://github.com/ossf/s2c2f/blob/main/specification/framework.md#appendix-relation-to-scitt
- This use case will be mostly focused on the policy / gatekeeper component and federation components of SCITT.
- 5.2.2: Registration Policies
- 7: Federation
- This use case is a specialization of (cross between) the following use cases from the Detailed Software Supply Chain Uses Cases for SCITT doc.
- 3.3: Security Analysis of a Software Product
- We'll cover OpenSSF Scorecard and other analysis mechanisms including meta static analysis / aggregation (example: GUAC).
- 3.4: Promotion of a Software Component by multiple entities
- We'll cover how these entities can leverage analysis mechanisms to achieve feature and bugfix equilibrium across the diverged environment.
- Future use cases could explore semantic patching to patch across functionally similar
- We'll cover how these entities can leverage analysis mechanisms to achieve feature and bugfix equilibrium across the diverged environment.
- 3.3: Security Analysis of a Software Product
- Alice builds a python Package
- Notary checks for receipts from needed sign offs
- In this example the SCITT instance the notary inserting into it have the same insert/sign policies (system context, dataflow, open architecture document, living threat model)
- Alice has two jobs, one which bulds a Python package, and other which runs SCITT in a TEE (perhaps with a redis service container to ease comms)
- She auths to job in TEE (SGX in this example) local SCITT via OIDC, the SCITT notary and ledger are unified in this example and a claim is inserted that she had a valid OIDC token for job_workflow_sha + repositoryUri + repository_id + job.. The enclave is then dumped at the end of the job so that it can be joined to an other transparency services. This enables decentralized hermetic builds via federation of transparency services (by grafting them into org sepcific registires ad-hoc via CD eventing of forge federation).
- Alice has two jobs, one which bulds a Python package, and other which runs SCITT in a TEE (perhaps with a redis service container to ease comms)
- The notary is what's verifying the OIDC token.
- We can runs-on an SGX machine to do that.
- Using confidential compute and attribute based trust we can authenticate to a usage policy, this is the place for on/off chain contract negotiation.
- Off chain would be whenever we have to enter a hermetic enviornment (IPVM).
- In this example the SCITT instance the notary inserting into it have the same insert/sign policies (system context, dataflow, open architecture document, living threat model)
- Activity Pub Actors for signoff, send to inbox requesting signoff (issue ops), they say okay I'll add this exception sign off for this use case /system context to SCITT
- Then policy violating system context collects all needed exception receipts, listens for their entry via listening to the SCITT ActivityPub stream, and then re-issues request for admissions along with exception receipts using overlay section of serialized system context object
- Notary checks for receipts from needed sign offs
2023-03-16: This will get reworked heavily as we align across https://codeberg.org/forgejo-contrib/discussions/issues/12, Rapunzel, and Alice
A methodology allowing organizations to nominate security contact points and policies via ActivityPub Actors. This allows for notifications to be federated of new lifecycle events. These lifecycle events might be VEX, SBOM, CSAF security advisory information, repository events, etc.
VEX documents should be aligned with the either the https://docs.oasis-open.org/csaf/csaf/v2.0/csaf-v2.0.html or OpenVEX specs: https://www.chainguard.dev/unchained/accelerate-vex-adoption-through-openvex. We can then communicate the IDs via ActivityPub like so.
When entities find security issues in source code, the correct channel to report security issues can be found if the repo has an RFC 9116 security.txt
file with a Contact
field. This contact field can be a URL which points to an ActivityPub Actor.
Via traversal of ActivityPub AcivityStream objects, reporters are enabled to discover reporting endpoints. Researchers are also enabled to receive up to date events by following declared ActivityPub Actors. When a researcher finds a vulnerability, they can submit their evidence to an eNotary (could be self notarized). The eNotary attests validity of the vuln and then replys to ActivityPub threads to facilite communication of valid vuln to upstream.
Imagine the following YAML as a directed graph whose upleveled pesudocode form is:
bob_vcs_repo:
security.txt:
Contact: https://activitypub.securitytxt.activitypub.example.org/bob
activitypub_service:
endpoint_url: https://activitypub.securitytxt.activitypub.example.org
actors:
bob:
attachment:
type: "PropertyValue"
name: "activitypubextensions"
value: "<a href=\"https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/1\" target=\"_blank\" rel=\"nofollow noopener noreferrer me\"><span class=\"invisible\">https://</span><span class=\"\">activitypub.securitytxt.activitypub.example.org/users/bob/statuses/1</span><span class=\"invisible\"></span></a>"
statuses:
- id: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/1"
content: "activitypubextensions"
replies:
- "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/2"
- id: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/2"
inReplyTo: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/1"
content: "activitypubsecuritytxt"
replies:
- "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/3"
- id: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/3"
inReplyTo: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/2"
content: "https://github.com/opencontainers/image-spec/raw/v1.0.1/schema/image-manifest-schema.json"
replies:
- "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/4"
- id: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/4"
inReplyTo: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/3"
content: "bob.registry.example.org/src_repo_name_contents_are_webhook_translated_to_vcs_push_manifest:sha256@babebabe"
- id: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/5"
inReplyTo: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/2"
content: "did:web:registry.example.com:policy-as-code:blocklist%40sha256%3Aaaaaaaaa"
replies:
- "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/6"
- id: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/6"
inReplyTo: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/5"
content: "did:web:registry.example.com:receipts:not_on_blocklist%40sha256%3Aaaaaaaaa"
alice:
statuses:
- id: "https://activitypub.securitytxt.activitypub.example.org/users/alice/statuses/1"
inReplyTo: "https://activitypub.securitytxt.activitypub.example.org/users/bob/statuses/4"
content: "alice.registry.example.org/vex_contents_are_openvex_from_scratch:sha256@babebabe"
- ActivityPub extensions for security.txt
- Can you put things in
@context
?, yes. Unsure if other servers will propagate events. - It this piggybacking within the content approach interoperable today, yes.
- Can you put things in
- Somewhere, something happened
- Bob tells Alice what happened
- Alice decides, do I care about what whappened? (the federated event)
- It's the triage process
- intel/cve-bin-tool#2639
- Take upstream policy (attached to incoming via
inReplyTo
and orreplies
, you'd have to decide if you want to dereference these, perhaps based on reputaion of propagator to reduce attack impact)
graph TD
subgraph lifecycle[BOM Component Trust Asserstion Generation Lifecycle]
subgraph lifecycle_background[Background]
watch_for_new_releases_on_repo[Scan: New Release / Tag on Repo]
end
subgraph lifecycle_submission[Submission]
new_release[Scan: New Release / Tag on Repo]
use_in_bom[Scan: Pulled into cached / downloaded PyPi index]
end
subgraph lifecycle_deployment[Scanning aka GUAC firewall pattern]
is_on_latest[Scan release/tag]
submit_to_allowlist[Submit repo and SHA of commit scanned as trust attestation: trusted/untrusted]
use_in_bom --> is_on_latest
new_release --> is_on_latest
is_on_latest --> submit_to_allowlist
end
subgraph lifecycle_transparency_service[Transparency Service]
add_to_ts[Add Trust Attestation to append only log]
list_ts_trusted[Index of attestations with result: Trusted]
add_to_ts --> list_ts_trusted
end
subgraph lifecycle_usage[Use in CI/CD]
check_ts[Check Transparency Service for all BOM items]
do_build_if_all_trusted[Run build if all trusted]
check_ts --> do_build_if_all_trusted
end
submit_to_allowlist --> add_to_ts
list_ts_trusted --> watch_for_new_releases_on_repo
watch_for_new_releases_on_repo --> new_release
list_ts_trusted --> check_ts
end
graph TD
subgraph home
h_prt[pull request target PRT flow]
subgraph home_tee
h_ts[transparency service]
end
h_guac[GUAC neo4j]
h_manifest[PEP 440 Manifest Change]
h_eval[Dependency Evaluation flow]
h_manifest -->|pull request submited triggers| h_prt
h_prt -->|source TCB protection ring admission control query<br>sync poll or waitformessage ActivityPub async| h_guac
h_guac -->|emit data for query not in graph| h_eval
h_eval -->|metric collection data<br>shouldi<br>home and new faraway| h_ts
h_ts -->|ActivityPub emit data added to graph<br>trigger ingest| h_guac
end
subgraph faraway
f_prt[pull request target PRT flow]
subgraph faraway_tee
f_ts[transparency service]
end
f_guac[GUAC neo4j]
f_manifest[PEP 440 Manifest Changed]
f_manifest -->|pull request submited triggers| f_prt
f_prt -->|source TCB protection ring admission control query<br>sync poll or waitformessage ActivityPub async| f_guac
f_ts -->|ActivityPub emit data added to graph<br>trigger ingest| f_guac
end
h_prt -->|admission control allowed dep change<br>create pull request to trigger downstream valdation<br>waitformessage and status check api<br>for downstream aka faraway results| f_manifest
f_guac -->|emit data for query not in graph| f_ts
h_ts -->|federate evaluated claims| f_ts
- A container image was created (
FROM
rebuild chain)- Bob's forge tells Alice's forge, here's the content address uri for the manifest just pushed
- Alice looks at the manifest, runs through all the packages she's maintaining in her forge
- She applies the threat model of each as an overlay when determining if she wants to propagate into her internal environment
- If any of these
- She applies the threat model of each as an overlay when determining if she wants to propagate into her internal environment
- Alice's downstream listener executes a system context to system context translation (grep: equilibrium, context-to-context)
- She executs the shim
- #1273
- It parses the content in alignment with the schema
- The shim already supports validation so we could actually just serialize the would be HTTP requests to files (same as staged when offline)
- https://github.com/intel/dffml/pull/1273/files#r794027710
- Could add activity style using this operation (function) as upstream, just copy paste and push to shim
- The shim already supports validation so we could actually just serialize the would be HTTP requests to files (same as staged when offline)
- https://github.com/tern-tools/tern#report-cyclonedxjson
- Upload resulting SBOM to registry
FROM scratch
style or via
- Upload resulting SBOM to registry
- She executs the shim
- A SBOM was published
- Bob's forge uploads an SBOM to the registry
- Alice's forge decides if she wants to propagate it (prioritizer, gatekeeper, umbrella)
- Alice looks at the manifest, runs through all the packages she's maintaining in her forge
- She applies the threat model of each as an overlay when determining if she wants to propagate into her internal environment
- If any of these use similar components as were mentioned in this SBOM, propagate
- Alice's listener receives the new SBOM event
- She uploads a manifest instance of a SLURM submit job spec to her registry
- A manifest instance of a IPVM/SLURM submit job was published to Alice's registry
- Bob's forge uploads an SBOM to the registry
- Alice's forge decides if she wants to propagate it (prioritizer, gatekeeper, umbrella)
- Alice looks at the manifest, runs through all the packages she's maintaining in her forge
- She applies the threat model of each as an overlay when determining if she wants to propagate into her internal environment
- If any of these use similar components as were mentioned in this SBOM, propagate
- Alice's listener within korifi receives the new SLURM submit job event
- She downloads the job contents from the manifest
FROM scratch
,results.yaml
extraction style tar pipe
- She executes the shim
- The next phase parser runs kaniko
- She downloads the job contents from the manifest
graph TD
subgraph bob[Bob's Cool Software]
actor[ActivityPub Actor - @ bob@ forge.mycoolsoftware.example.com]
actor_attachment[Attachment PropertyValue activitypubsecuritytxt]
activitypubsecuritytxt_root_post[activitypubsecuritytxt root post]
activitypubsecuritytxt_vcs_push[vcs.push root post]
activitypubsecuritytxt_vcs_push_content[vcs.push content - content address of manifest instance in registry]
actor --> actor_attachment
actor_attachment -->|Link| activitypubsecuritytxt_root_post
activitypubsecuritytxt_vcs_push -->|inReplyTo| activitypubsecuritytxt_root_post
activitypubsecuritytxt_vcs_push_content -->|inReplyTo| activitypubsecuritytxt_vcs_push
end
subgraph alice[Alice]
alice_shouldi_contribute[Static Analysis Result] -->|inReplyTo| activitypubsecuritytxt_vcs_push_content
end
- So no matter where you're executing, all the reporting and eventing is the same, because we are loosely coupled.
- We can do
fromjson
in jq or we can do more advanced xargs chaining on the websocket for ad-hox dev work - We can shoot from the activitypub inbox receiver to a message queue for integration with existing celery
- This way we sidestep all rate limiting except for when we have to preform write events to GitHub
- Otherwise we always read GitHub data from cypher queries over the reboardcast data
- We can also have listeners which reboardcast the resolved contents of content address style broadcast data (the top level, so if this sees a container image uri broadcast, it would be pulling it down and maybe rebroadcasting the
results.yaml
or whatever is they transform needed to rebroadcast that data. - This is our onramp into the linked data space, eventually KERI for backing comms security
- We can also have listeners which reboardcast the resolved contents of content address style broadcast data (the top level, so if this sees a container image uri broadcast, it would be pulling it down and maybe rebroadcasting the
- We can do
- We'll implement an InputNetwork fro DFFML so that every fucntion/operation automaticlly gets added into the noe/GUAC graph on execution (triggering eventing via SBOM of query).
schema/alice/shouldi/contribute/github-com-omnilib-aiosqlite.json
{
"@context": "https://github.com/intel/dffml/raw/alice/schema/schema/alice/shouldi/contribute/0.0.0.schema.json",
"repo_url": "https://github.com/omnilib/aiosqlite"
}
- As container build
schema/image/container/build/alice-shouldi-contribute-results-github-com-omnilib-aiosqlite.json
{
"@context": "https://github.com/intel/dffml/raw/alice/schema/github/actions/build/images/containers/0.0.0.schema.json",
"include": [
{
"branch": "alice",
"build_args": "[[\"REPO_URL\", \"https://github.com/omnilib/aiosqlite\"]]",
"commit": "ca92bfae5092bce908b70f6b5e0afbe242ce7a5b",
"dockerfile": "entities/alice/scripts/alice-shouldi-contribute-results.Dockerfile",
"image_name": "alice-shouldi-contribute-results-github-com-omnilib-aiosqlite",
"owner": "intel",
"repository": "dffml"
}
]
}
- Open Source scanning flow
- https://github.com/ossf/s2c2f/blob/main/specification/framework.md#appendix-relation-to-scitt
- Upload manifest to registry
- Federation event (send to follower /inbox)
- content:
https://github.com/opencontainers/image-spec/raw/v1.0.1/schema/image-manifest-schema.json
inReplyTo: activitypub extensions for security.txt post URL for contentactivitypubsecuritytxt
- content: container image uri uploaded
inReplyTo: activitypub extensions for security.txt post URL for content
https://github.com/opencontainers/image-spec/raw/v1.0.1/schema/image-manifest-schema.json
- content:
- Federation event (send to follower /inbox)
- Downstream listener (aka delve into config dict)
- Federation event (send to follower /inbox)
- content:
https://github.com/intel/dffml/raw/alice/schema/github/actions/build/images/containers/0.0.0.schema.json
inReplyTo: activitypub extensions for security.txt post URL for contentactivitypubsecuritytxt
- content:
<extracted content(?)>
inReplyTo: activitypub extensions for security.txt post URL for contenthttps://github.com/intel/dffml/raw/alice/schema/github/actions/build/images/containers/0.0.0.schema.json
- content:
- Federation event (send to follower /inbox)
- Downstream listener
- Republish watched
inReplyTo
schema into job/message queue- RabbitMQ
- Message queue delivers to worker nodes
- Kaniko job waiting for celery queue for image to build
- Exit after rebuild and have orchestration manage respawn
- Kaniko job waiting for celery queue for image to build
- Republish watched
graph LR
subgraph vcs_source[Version Controled Software]
subgraph dffml_vcs_source[dffml.git]
subgraph dffml_vcs_source_security_txt[security.txt]
dffml_vcs_source_security_txt_contact[Contact: https://example.org/dffml]
end
subgraph dffml_vcs_source_dockerfile[dffml.Dockerfile]
dffml_vcs_source_dockerfile_from_base[FROM upstream as dffml]
end
subgraph dffml_vcs_source_dockerfile_example[dffml.example.Dockerfile]
dffml_vcs_source_dockerfile_example_from_base[FROM dffml @ sha:latest]
end
subgraph vcs_source_alice[dffml.git/entities/alice]
subgraph alice_vcs_source_security_txt[security.txt]
alice_vcs_source_security_txt_contact[Contact: https://example.org/alice]
end
subgraph alice_vcs_source_dockerfile[alice.Dockerfile]
alice_vcs_source_dockerfile_from_base[FROM dffml @ sha:latest]
end
subgraph alice_vcs_source_dockerfile_shouldi_contribute[alice_shouldi_contribute.Dockerfile]
alice_vcs_source_dockerfile_shouldi_contribute_from_base[FROM alice @ sha:latest]
subgraph alice_shouldi_contribute[alice shoulid contribute -keys ARG_REPO_URL]
alice_shouldi_contribute_git_clone[git clone ...]
alice_shouldi_contribute_read_security_txt[grep Contact: security.txt]
alice_shouldi_contribute_result[Static Analysis Result]
alice_shouldi_contribute_git_clone --> alice_shouldi_contribute_read_security_txt
dffml_vcs_source_security_txt_contact --> alice_shouldi_contribute_read_security_txt
alice_shouldi_contribute_read_security_txt --> alice_shouldi_contribute_result
end
end
end
end
end
subgraph schema[Manifest ADRs]
subgraph manifest_build_images_contianers[Build Image Container]
manifest_build_images_contianers_intent[README.md/THREATS.md]
manifest_build_images_contianers_schema[1.0.0.schema.json]
end
end
subgraph manifest_instances[Manifest Instances]
alice_manifest_build_images_contianers_alice_shouldi_contribute
end
subgraph transparency_logs[Transparency Logs]
dffml_scitt[dffml.scitt.example.org]
alice_scitt[alice.scitt.example.org]
end
subgraph factory[Secure Software Factories aka Federated Forge]
subgraph build_images_contianers[build_images_contianers.yml]
end
subgraph factory_container_image_registries[Container Image Registry https://oras.land]
subgraph dffml_factory_container_image_registries_project[DFFML Images]
dffml_container_image[dffml:latest]
end
subgraph alice_factory_container_image_registries_project[Alice Images]
alice_container_image[alice:latest]
alice_shouldi_contribute_scan_results[shouldicontribute @ sha384:babebabe]
end
end
build_images_contianers --> dffml_scitt
build_images_contianers --> alice_scitt
end
subgraph protocol_knowledge_graph_activity_pub[ActivityPub]
subgraph ActivityPubExtensionsForSecurityTXT[activitypub extensions for security.txt]
subgraph dffml_security_txt_contact[dffml.git/security.txt:Contact]
dffml_actor[ActivityPub Actor - @ dffml @ example.org]
dffml_actor_attachment[Attachment PropertyValue activitypubsecuritytxt]
dffml_activitypubsecuritytxt_root_post[activitypubsecuritytxt root post]
dffml_activitypubsecuritytxt_vcs_push[vcs.push root post]
dffml_activitypubsecuritytxt_vcs_push_content[vcs.push content - content address of manifest instance in registry]
dffml_actor --> dffml_dffml_actor_attachment
dffml_actor_attachment -->|Link| dffml_activitypubsecuritytxt_root_post
dffml_activitypubsecuritytxt_vcs_push -->|inReplyTo| dffml_activitypubsecuritytxt_root_post
dffml_activitypubsecuritytxt_vcs_push_content -->|inReplyTo| dffml_activitypubsecuritytxt_vcs_push
end
subgraph alice_security_txt_contact[dffml.git/entites/alice/security.txt:Contact]
alice_actor[ActivityPub Actor - @ alice @ example.org]
alice_actor_attachment[Attachment PropertyValue activitypubsecuritytxt]
alice_activitypubsecuritytxt_root_post[activitypubsecuritytxt root post]
alice_activitypubsecuritytxt_vcs_push[vcs.push root post]
alice_activitypubsecuritytxt_vcs_push_content[vcs.push content - content address of manifest instance in registry]
alice_actor --> alice_actor_attachment
alice_actor_attachment -->|Link| alice_activitypubsecuritytxt_root_post
alice_activitypubsecuritytxt_vcs_push -->|inReplyTo| alice_activitypubsecuritytxt_root_post
alice_activitypubsecuritytxt_vcs_push_content -->|inReplyTo| alice_activitypubsecuritytxt_vcs_push
end
end
alice_actor -->|follow| dffml_actor
end
subgraph render_knowledge_graph_agora[Agora]
end
alice_vcs_source_dockerfile_shouldi_contribute
dffml_vcs_source_security_txt_contact --> dffml_actor
alice_vcs_source_security_txt_contact --> alice_actor
alice_shouldi_contribute_result --> alice_shouldi_contribute_scan_results
alice_shouldi_contribute_scan_results --> |inReplyTo| dffml_vcs_source_dockerfile_example_from_base
dffml_container_image --> dffml_vcs_source_dockerfile_example_from_base
alice_container_image --> alice_vcs_source_dockerfile_example_from_base
dffml_vcs_source_dockerfile_example_from_base --> dffml_activitypubsecuritytxt_vcs_push
dffml_activitypubsecuritytxt_vcs_push --> build_images_contianers_trigger
alice_vcs_source_dockerfile_example_from_base --> alice_activitypubsecuritytxt_vcs_push
alice_shouldi_contribute
{
"@context": [
"https://www.w3.org/ns/activitystreams",
"https://w3id.org/security/v1",
],
"id": "https://mastodon.social/users/alice",
"type": "Person",
"following": "https://mastodon.social/users/alice/following",
"followers": "https://mastodon.social/users/alice/followers",
"inbox": "https://mastodon.social/users/alice/inbox",
"outbox": "https://mastodon.social/users/alice/outbox",
"featured": "https://mastodon.social/users/alice/collections/featured",
"featuredTags": "https://mastodon.social/users/alice/collections/tags",
"preferredUsername": "alice",
"name": "Alice",
"summary": "An ActivityPub Actor",
"url": "https://mastodon.social/@alice",
"publicKey": {
"id": "https://mastodon.social/users/alice#main-key",
"owner": "https://mastodon.social/users/alice",
"publicKeyPem": "-----BEGIN PUBLIC KEY-----\nMIIBIjANBgk\n-----END PUBLIC KEY-----\n"
},
"attachment": [
{
"type": "PropertyValue",
"name": "activitypubextensions",
"value": "<a href=\"https://mastodon.social/users/alice/statuses/1\" target=\"_blank\" rel=\"nofollow noopener noreferrer me\"><span class=\"invisible\">https://</span><span class=\"\">mastodon.social/users/alice/statuses/1</span><span class=\"invisible\"></span></a>"
}
],
"endpoints": {
"sharedInbox": "https://mastodon.social/inbox"
}
}
{
"@context": [
"https://www.w3.org/ns/activitystreams"
],
"id": "https://mastodon.social/users/alice/statuses/1",
"type": "Note",
"summary": null,
"inReplyTo": null,
"published": "2022-11-11T04:40:17Z",
"url": "https://mastodon.social/@alice/1",
"attributedTo": "https://mastodon.social/users/alice",
"to": [
"https://www.w3.org/ns/activitystreams#Public"
],
"cc": [
"https://mastodon.social/users/alice/followers"
],
"sensitive": false,
"content": "activitypubextensions",
"updated": "2022-11-11T04:42:27Z",
"attachment": [],
"replies": {
"id": "https://mastodon.social/users/alice/statuses/1/replies",
"type": "Collection",
"first": {
"type": "CollectionPage",
"next": "https://mastodon.social/users/alice/statuses/1/replies?min_id=1&page=true",
"partOf": "https://mastodon.social/users/alice/statuses/1/replies",
"items": [
"https://mastodon.social/users/alice/statuses/2"
]
}
}
}
- TODO for root nodes, should we be using a manifest instance?
{
"@context": [
"https://www.w3.org/ns/activitystreams"
],
"id": "https://mastodon.social/users/alice/statuses/2",
"type": "Note",
"summary": null,
"inReplyTo": "https://mastodon.social/users/alice/statuses/1",
"published": "2022-11-11T04:40:17Z",
"url": "https://mastodon.social/@alice/2",
"attributedTo": "https://mastodon.social/users/alice",
"to": [
"https://www.w3.org/ns/activitystreams#Public"
],
"cc": [
"https://mastodon.social/users/alice/followers"
],
"sensitive": false,
"content": "activitypubsecuritytxt",
"updated": "2022-11-11T04:42:27Z",
"attachment": [],
"replies": {
"id": "https://mastodon.social/users/alice/statuses/2/replies",
"type": "Collection",
"first": {
"type": "CollectionPage",
"next": "https://mastodon.social/users/alice/statuses/2/replies?min_id=2&page=true",
"partOf": "https://mastodon.social/users/alice/statuses/2/replies",
"items": [
"https://mastodon.social/users/alice/statuses/3"
]
}
}
}
{
"@context": [
"https://www.w3.org/ns/activitystreams"
],
"id": "https://mastodon.social/users/alice/statuses/3",
"type": "Note",
"summary": null,
"inReplyTo": "https://mastodon.social/users/alice/statuses/2",
"published": "2022-11-11T04:40:17Z",
"url": "https://mastodon.social/@alice/3",
"attributedTo": "https://mastodon.social/users/alice",
"to": [
"https://www.w3.org/ns/activitystreams#Public"
],
"cc": [
"https://mastodon.social/users/alice/followers"
],
"sensitive": false,
"content": "https://github.com/opencontainers/image-spec/raw/v1.0.1/schema/image-manifest-schema.json",
"updated": "2022-11-11T04:42:27Z",
"attachment": [],
"replies": {
"id": "https://mastodon.social/users/alice/statuses/3/replies",
"type": "Collection",
"first": {
"type": "CollectionPage",
"next": "https://mastodon.social/users/alice/statuses/3/replies?min_id=3&page=true",
"partOf": "https://mastodon.social/users/alice/statuses/3/replies",
"items": [
"https://mastodon.social/users/alice/statuses/4"
]
}
}
}
{
"@context": [
"https://www.w3.org/ns/activitystreams"
],
"id": "https://mastodon.social/users/alice/statuses/4",
"type": "Note",
"summary": null,
"inReplyTo": "https://mastodon.social/users/alice/statuses/3",
"published": "2022-11-11T04:54:56Z",
"url": "https://mastodon.social/@alice/4",
"attributedTo": "https://mastodon.social/users/alice",
"to": [
"https://www.w3.org/ns/activitystreams#Public"
],
"cc": [
"https://mastodon.social/users/alice/followers"
],
"sensitive": false,
"content": "registry.example.org/vex:sha256@babebabe",
"attachment": [],
"tag": [],
"replies": {
"id": "https://mastodon.social/users/alice/statuses/4/replies",
"type": "Collection",
"first": {
"type": "CollectionPage",
"next": "https://mastodon.social/users/alice/statuses/4/replies?only_other_accounts=true&page=true",
"partOf": "https://mastodon.social/users/alice/statuses/4/replies",
"items": []
}
}
}
- Now we want to translate to OpenVEX and have the content addresses of the signature for the post
statements.impact_statement
is Webhook payload object with SCITT enchancements- https://scitt.io/distributing-with-oci-scitt.html
- Registry content addresses contains granular results (webhook payload,
alice shouldi contribute
, etc.)- Then our webhook watch on the registry publishes the replys
- Or we update an example container with a pinned sha on the
FROM
- By watching the
push@vcs
(Version Control System) for the file
- By watching the
- We should upload the VEX without the
@id
to the registry, then use that ID as the VEX@id
when wecreatePost()
.- Or better yet just have it do a kontain.me lightwieght proxy from the registry object
{
"@context": "https://openvex.dev/ns",
"@id": "https://mastodon.social/users/alice/statuses/vex-sha256@feedface",
"author": "GitHub Actions <[email protected]>",
"role": "GitHub Actions",
"timestamp": "2023-02-02T14:24:00.000000000-07:00",
"version": "1",
"statements": [
{
"vulnerability": "vex-vcspush-sha256@feedface",
"products": [
"pkg:github/intel/dffml@ddb32a4e65b0d79c7561ce2bdde16d963c8abde1"
],
"status": "not_affected",
"justification": "vulnerable_code_not_in_execute_path"
"impact_statement": "registry.example.org/vcspush:sha256@feedface",
}
]
}
- TODO https://docs.oasis-open.org/csaf/csaf/v2.0/csaf-v2.0.html example
- We can now watch for events
- TODO Update the following with the
xargs
call to https://github.com/genuinetools/reg (or equivilant) to do the double hop download of the VEX (without the@id
right now because there is no proxy, just add on download), and then the subsquent manifest download of the vsspush webhook payload, then finally do the.commits[].modified[]
selection$ websocat --exit-on-eof --basic-auth alice:$(cat ../password) ws://localhost:8000/listen/websocket | jq --unbuffered -r .object.id | xargs -l -I '{}' -- sh -c "curl -sfL '{}' | jq -r" &
- jakelazaroff/activitypub-starter-kit#2
- TODO Update the following with the
$ curl -sfL https://vcs.example.org/push/outbox | jq --unbuffered -r '.orderedItems[].object.content' | grep stream_of | grep modified | jq -r --unbuffered '.commits[].modified[]'
Dockerfile
- Example use cases
- DFFML 2nd party downstream rebuilds
- https://github.com/intel/dffml/pull/1061/files#diff-c7d7828822f15922ed830bb6f3148edc97c291c809836b1a1808165d36bd8c9dR225-R229
- Rebuild upstream container when we get an VEX (via AcivityPub) from upstream saying that any of the files we want to watch have changed
- At first we will just watch all files within the downstream container build workflow
on.workflow_dispatch && on.push.paths: ["https://github.com/intel/dffml.git#branch=main/*"]
- Later we will watch for the example container with the pinned version
on.workflow_dispatch && on.push.paths: ["https://github.com/intel/dffml.git#branch=main/dffml/util/skel/common/Dockerfile"]
dffml/util/skel/common/Dockerfile
FROM registry.dffml.org/dffml:sha256@babebabe
- At first we will just watch all files within the downstream container build workflow
- DFFML 2nd party downstream rebuilds
Actors can be spun up ad-hoc, mirrors decentralized nature of OSS development.
Enables projects to update based on policy.
sequenceDiagram
SupportLevelOneSecondPartyPluginPullRequest->>+Webhook:
Webhook->>+ActivityPub:
ActivityPub->>+SupportLevelOneSecondPartyWatcher:
RequiredCIJobPassDownstreamSupportLevelOneSecondPartyPlugin->>+Webhook:
Webhook->>+ActivityPub:
SupportLevelOneSecondPartyPlugin->>+Webhook:
DFFML->>+Webhook:
Webhook->>+ActivityPub:
SupportLevelOneSecondPartyPluginPullRequest->>+Webhook:
Webhook->>+ActivityPub:
ActivityPub->>+DFFMLTutorialPullRequestBumpPin:
ExampleDockerFileWithPinnedVersion->>+ContainerBuild:
ContainerBuild->>+ExampleDockerFileWithPinnedVersion:
ExampleDockerFileWithPinnedVersion->>+Webhook:
ActivityPub->>+DFFMLTutorialPullRequestBumpPin:
DFFMLTutorialPullRequestBumpPin->>+Webhook:
Upstream of following mermaid: https://github.com/intel/dffml/tree/alice/docs/tutorials/rolling_alice/0000_architecting_alice#what-is-alice
graph BT
subgraph Alice[Alice the Entity]
subgraph compute[Compute]
Web5[Web 5]
KCP
CI_CD[CI/CD]
end
subgraph soul[Strategic Plans and Principles]
Threat_Modeling[Threat Modeling]
Debug
end
subgraph collector[Collector]
subgraph dynamic_analysis[Dynamic Analysis]
policy[policy.yml]
sandbox_policy_generator[Adaptive Sandboxing]
end
subgraph static_analysis[Static Analysis]
cve_bin_tool[CVE Binary Tool]
SBOM
end
end
Open_Architecture
Open_Architecture[Alice the Open Architecture]
snapshot_system_context[Alice the Overlay<br>Snapshot of System Context]
orchestartor[Orchestartor]
Open_Architecture --> Threat_Modeling
Open_Architecture --> Debug
Threat_Modeling --> orchestartor
Debug --> orchestartor
orchestartor --> KCP
orchestartor --> Web5
orchestartor --> CI_CD
CI_CD --> snapshot_system_context
KCP --> snapshot_system_context
Web5 --> snapshot_system_context
snapshot_system_context --> sandbox_policy_generator
snapshot_system_context --> cve_bin_tool
sandbox_policy_generator --> policy --> Open_Architecture
cve_bin_tool --> SBOM --> Open_Architecture
cve_bin_tool --> VEX -->|Trigger validation run of mitigation suggestion| orchestartor
policy -->|Check if policy says out of scope<br>client vs. server usage| VEX
end
- Just FYI, have been playing with the idea of using security.txt contact as an AcivityPub Actor to advertise things such as delegate Actors for various purposes. For example, list via attachments actors which publish content addresses of an orgs SBOMs This would enable leveraging ActivityPub as a means for definition and broadcast for entities delegated to various roles. We could do the same for the 3rd parties to advertise what actors are within which roles, aka are authorized to say this thing is FIPs certified. We could then attach SCITT receipts to these: intel/dffml#1406
- The SCITT registry then becomes the quick lookup path (analogously database view) to verify this. This way end users don't have to traverse the full Knowledge Graph (Activity Pub in this case). Receipt we care about for verification would be is this
inReplyTo
DAG hop path valid, aka isdid:merkle
in SCITT. - Can have a thread linked in attachments for manifests, can discover from there
- Can watch for replies and execute jobs based off listening for manifest instances
inReplyTo
to the manifest.- Post content addresses of manifest existing in oras.land (a container "image" registry)
- Can watch for replies and execute jobs based off listening for manifest instances
- https://github.com/WebOfTrustInfo/rwot11-the-hague/blob/master/advance-readings/Enhancing_DID_Privacy_through_shared_Credentials.md
- https://github.com/WebOfTrustInfo/rwot11-the-hague/blob/master/draft-documents/did-merkle.md
- The SCITT registry then becomes the quick lookup path (analogously database view) to verify this. This way end users don't have to traverse the full Knowledge Graph (Activity Pub in this case). Receipt we care about for verification would be is this
- Looks like we can have four attachments, we can make one link to a post as an attachment, then replies to that to build more trees of data
- https://policymaker.disclose.io/policymaker/introduction
- References
- RFC9116: https://securitytxt.org/
- ietf-scitt#14
- openvex/spec#9
- https://forum.forgefriends.org/t/about-the-friendly-forge-format-f3/681
-
ForgeFed is an ActivityPub extension. ActivityPub is an actor-model based protocol for federation of web services and applications.
-
- https://mastodon.social/@[email protected]/109784681116604896
-
meanwhile at work, a thing i've been working on for the past few months has dropped: https://www.chainguard.dev/unchained/accelerate-vex-adoption-through-openvex it's basically like ActivityStreams, but for security vulnerability data sharing. with a little bit of work, we can lift up to something more like ActivityPub for real-time collaboration, a blog is forthcoming about it.
-
- aka the Manifest Transport ADR
- Associated Alice tutorial: Rolling Alice: Architecting Alice: Stream of Consiousness
- https://social.treehouse.systems/@ariadne/109808644259234008
- We'll want to align with Ariadne's Rapunzel
- https://social.treehouse.systems/@ariadne/109808644259234008
- Associated Alice tutorial: Rolling Alice: Architecting Alice: Stream of Consiousness
- TODO
- OIDC to keypair to post replys (fulcio?)
- Or just the noterizing proxy
- OIDC to keypair to post replys (fulcio?)
- https://codeberg.org/fediverse/fep
- service: sw: src: change: notify: Service to facilitate poly repo pull model dev tooling: activitypubsecuritytxt
- Reproduced below
- We seek interop with Aradine's Rapunzel
- The Agora: a Knowledge Commons
- https://docs.google.com/document/d/1DXJRDh9Ss5VCRBi3oirDw9d7yjn3H2hMqfN2ETTyjIc/edit#
- We seek interop with the Agora
- Scratch work upstream: intel/dffml#1406
- Don't forget about KCP
- Run some live ones in https://github.com/cloudfoundry/korifi via
dffml-service-http
- Demo similar job URL hash as registry tag based addressing of results within registry
- Enable sending of AcivityPub events directly (later) or indirectly via proxy nodes (first, activitypub starter kit.
- https://linkeddatafragments.org/
- http://query.linkeddatafragments.org/#query=&resultsToTree=false&queryFormat=graphql
- https://gist.github.com/rubensworks/9d6eccce996317677d71944ed1087ea6
- https://github.com/comunica/jQuery-Widget.js/blob/master/config/config-default.json
- We need to turn the stream into something we can query using cypher or graphql-ld
- https://swordapp.github.io/swordv3/swordv3.html
- https://oras.land/blog/gatekeeper-policies-as-oci-image/
- https://github.com/project-zot/zot
- Okay if we can make the KERI SCITT instance use the OCI upload/download spec and then align the telemetry and registry federation protocols
- Look into existing registry federation protocol if exists
- https://s3hh.wordpress.com/2022/10/27/oci-based-linux/
- Similar goals to OS DecentrAlice
- https://github.com/project-machine/mos/releases/tag/0.0.7
- https://github.com/opencontainers/distribution-spec/blob/main/spec.md#endpoints
- opencontainers/distribution-spec#388
- Have we thought about federation protocols / APIs? To enable registries to propagate uploaded content within a network of registries? Looking to come up to speed on any existing discussion if that's been touched on. Thank you!
- References
- https://github.com/opencontainers/distribution-spec/blob/main/spec.md#endpoints
- Looked here for relevant paths here but not seeing anything that looks like it's for notifications / inbox style eventing
- https://github.com/sapcc/keppel
- ietf-scitt#14
- Hoping we can align to similar federation protocols across transparency services and container registries so event stream consumers can work with the same protocol for each (ActivityStreams/Pub?)
- https://github.com/opencontainers/distribution-spec/blob/main/spec.md#endpoints
- https://conformance.opencontainers.org/
- https://vsoch.github.io/django-oci/docs/getting-started/auth
- https://vsoch.github.io/django-oci/docs/getting-started/testing
- opencontainers/distribution-spec#110 (comment)
- https://github.com/sapcc/keppel
- https://github.com/sapcc/keppel/blob/master/docs/api-spec.md#post-keppelv1authpeering
- Looks like they have their own spec for federation, maybe we can implement with ActivityPub?
- Maybe we can leverage the existing APIs similar to the /admin endpoint and just add in the activitypub endpoints for activitystreams / linked data notifications
- Looks like they have their own spec for federation, maybe we can implement with ActivityPub?
- https://github.com/sapcc/keppel/blob/master/docs/example-policy.yaml
- We can take one manifest and make it into another one for execution via a different mechanism
- Similar to the CLI overlays
- This is also similar to how we can decouple TODO logging from content for
alice please log todos
- Operation to generate TODO body
- Operation for logging the TODO (write to GitHub)
- Similar to a mutation of the propagated event into something context local relevant
- Yes this vuln affects due to instance policy relevant threat model overlays or not
- https://github.com/opencontainers/image-spec/blob/main/artifact.md
- Manifest for CLI command
- The Open Architecture (Alice) sits at the interesction of CI/CD, Security, and AI/ML.
- We metion Alice here as a follow on who's development sees this use case as critical
- Think cross between review system (SCITT as the proof, TDB on identity preknown at this point, OpenSSF members stream 8 vuln sharing CCF ledger)
- https://scitt.io/distributing-with-oci-scitt.html
- https://www.darkreading.com/dr-tech/cybersecurity-nutrition-labels-still-a-work-in-progress
- https://www.whitehouse.gov/briefing-room/statements-releases/2022/10/20/statement-by-nsc-spokesperson-adrienne-watson-on-the-biden-harris-administrations-effort-to-secure-household-internet-enabled-devices/
-
Yesterday, the White House convened leaders from the private sector, academic institutions, and the U.S. Government to advance a national cybersecurity labeling program for Internet-of-Things (IoT) devices. The Biden-Harris Administration has made it a priority to strengthen our nation’s cybersecurity, and a key part of that effort is ensuring the devices that have become a commonplace in the average American household – like baby monitors or smart home appliances – are protected from cyber threats. A labeling program to secure such devices would provide American consumers with the peace of mind that the technology being brought into their homes is safe, and incentivize manufacturers to meet higher cybersecurity standards and retailers to market secure devices.
Yesterday’s dialogue focused on how to best implement a national cybersecurity labeling program, drive improved security standards for Internet-enabled devices, and generate a globally recognized label. Government and industry leaders discussed the importance of a trusted program to increase security across consumer devices that connect to the Internet by equipping devices with easily recognized labels to help consumers make more informed cybersecurity choices (e.g., an “EnergyStar” for cyber). These conversations build on the foundational work that has been pioneered by the private sector and the National Institute of Standards and Technology (NIST) to help build more secure Internet-connected devices. It also follows President Biden’s Executive Order on Improving the Nation’s Cybersecurity, which highlighted the need for improved IoT security and tasked NIST, in partnership with the Federal Trade Commission, to advance improved cybersecurity standards and standardized product labels for these devices.
-
- https://www.whitehouse.gov/briefing-room/statements-releases/2022/10/20/statement-by-nsc-spokesperson-adrienne-watson-on-the-biden-harris-administrations-effort-to-secure-household-internet-enabled-devices/
- https://csrc.nist.gov/publications/detail/white-paper/2022/11/09/implementing-a-risk-based-approach-to-devsecops/final
-
DevOps brings together software development and operations to shorten development cycles, allow organizations to be agile, and maintain the pace of innovation while taking advantage of cloud-native technology and practices. Industry and government have fully embraced and are rapidly implementing these practices to develop and deploy software in operational environments, often without a full understanding and consideration of security. Also, most software today relies on one or more third-party components, yet organizations often have little or no visibility into and understanding of how these components are developed, integrated, deployed, and maintained, as well as the practices used to ensure the components’ security. To help improve the security of DevOps practices, the NCCoE is planning a DevSecOps project that will focus initially on developing and documenting an applied risk-based approach and recommendations for secure DevOps and software supply chain practices consistent with the Secure Software Development Framework (SSDF), Cybersecurity Supply Chain Risk Management (C-SCRM), and other NIST, government, and industry guidance. This project will apply these DevSecOps practices in proof-of-concept use case scenarios that will each be specific to a technology, programming language, and industry sector. Both closed source (proprietary) and open source technology will be used to demonstrate the use cases. This project will result in a freely available NIST Cybersecurity Practice Guide.
-
- https://www.intel.com/content/www/us/en/newsroom/news/2022-intel-innovation-day-2-livestream-replay.html#gs.djq36o
- Similar to the software labeling, with Alice we are trying to cross these streams
- Datasheets for Datasets
- https://arxiv.org/abs/1803.09010
-
The machine learning community currently has no standardized process for documenting datasets, which can lead to severe consequences in high-stakes domains. To address this gap, we propose datasheets for datasets. In the electronics industry, every component, no matter how simple or complex, is accompanied with a datasheet that describes its operating characteristics, test results, recommended uses, and other information. By analogy, we propose that every dataset be accompanied with a datasheet that documents its motivation, composition, collection process, recommended uses, and so on. Datasheets for datasets will facilitate better communication between dataset creators and dataset consumers, and encourage the machine learning community to prioritize transparency and accountability.
- Datasheets for Datasets
- Similar to the software labeling, with Alice we are trying to cross these streams
Side from Andrew Ng's Intel Innovation 2022 Luminary Keynote Source: https://www.intel.com/content/www/us/en/newsroom/news/2022-intel-innovation-day-2-livestream-replay.html#gs.iex8mr
- Possible alignment with Andrew's "Data-Centric AI"
- is the discipline of systematically engineering the data used to build an AI system
- This is what we're doing with Alice
- Possible alignment with Andrew's "The iterative process of ML development"
- https://github.com/intel/dffml/tree/alice/docs/tutorials/rolling_alice/0000_architecting_alice#entity-analysis-trinity
- Intent / Train model
- Establish correlations between threat model intent and collected data / errors (telemetry or static analysis, policy, failures)
- Dynamic analysis / Improve data
- We tweak the code to make it do different things to see different data. The application of overlays. Think over time.
- Static / Error analysis
- There might be async debug initiated here but this maps pretty nicely conceptually since we'd think of this as a static process, we already have some errors to analyze if we're at this step.
Gif of Alice on roller skates throwing a bowling ball which is a software vuln, strike, she frontflips throwing knife style throws the pins into pull requests. We zoom out and see her just doing this over and over again around the Entity Analysis Trinity. Intent/LTM is where the throwing board is. Bowling alley is static analysis and the end of the bowling ally where she frontflips over (through hoop of CI/CD fire?) is where she pics up the pins and throws them as pull request (titles and numbers maybe, pulls/1401 style maybe?) knives into the board at the top which is the LTM and codebase. Then from top, LTM to static analysis where bowling alley starts shes in the lab, cooking up the vuln or maybe out looking for it. Or maybe refactoring after pull requests!
Mike from OpenSSF has been thinking about SCITT as a schema and rules on how one would assert facts, weither it's confidential compute or traditional permissions is impelmenetation details.
- TODO
- Going with Authenticated Chained Data Containers and KERI Watchers as an implmentation of SCITT for this
- It's basically a decentralized pubsub event notification methodology that can be done over ACDC piggybacking on ActivityPub as layer 7.
- Event data lives "off chain" in a container registry secured via existing transparency service based methods (KERI, SCITT, SigStore), where the chain is the network of keys involved for a train of thoughts comms between entities. Since we transmit ActivityPub over KERI, the graph of our supply chain data we are sharing can be shared with trusted actors who agree not to be duplicitous, and who's KERI keys can be tied back to TEEs so that we can confirm they are running software that doesn't intend (via ML-based, Alice, analysis) to be duplicitous. We can now have our trusted computing based for decentralized compute, aka CI/CD pipelines delivering across project trust boundries.
- Go over
.github/workflows/alice_shouldi_contribute.yml
which is called as reusable workflow using SLSA demos as conceptual upstream.- This gives us metric collection with overlays applied to input network which log data provenance later used via policy referenced in a Living Threat Model of a downstream end user facing application. The use case document will cover how we effectivly self notarize the data provenance assisted by the OIDC token issused to the workflow from GitHub.
- Align with Software Supply Chain Artifact Examples use case
- As CI/CD velocity increases, we approach a similar threat model to distributed compute, address here.
- As consumer, how can I discover that fact and trust that it's accruate
- Could imagine a world where things like Scorecard express the data as as SCITT attestation
- You go and query that (source code data mining data or vuln data) store (via DID resolution)
and you say tell me everythig you know about foo and you get it all back.
- Addresses OpenSSF/Christine's: Looking at trying to connect all the different data sources
- Until we have an implementation with Web5 (https://github.com/TBD54566975) that's at at least beta, we could expore what that looks like.
- Open Architecture: need to track usages /
reuse
of contextsictx
,nctx
, etc. with something predeclared, aka at runtime if yourOperation
data structure doesn't allowlist your usage of it you can pass it to a subflow for reuse. This allows us to use the format within our orchrestration and for static analysis because we can use this same format to describe the trust boundry proeprties that other domain sepcific represenatations of architecture have, for instance we could if we were doing and Open Architecture (OA) Intermediate Representation (IR) for and ELF file we might note that the input network context is not reused from the top level system context. Where as if we did an OA IR for Python code we would say that the input network is reused from the top level system context (it has access to that memory region, whereas when you launch and ELF you look access to the parents memory region, typically).
- Going with Authenticated Chained Data Containers and KERI Watchers as an implmentation of SCITT for this
- References
- https://scitt.io/distributing-with-oci-registries.html
- https://github.com/intel/dffml/blob/alice/docs/arch/0009-Open-Architecture.rst
- https://github.com/intel/dffml/blob/alice/docs/arch/0008-Manifest.md
- https://www.cisa.gov/sites/default/files/publications/VEX_Use_Cases_Aprill2022.pdf
- 2022-09-29 IETF SCITT Technical Meeting
- https://github.com/transmute-industries/did-eqt/blob/main/docs/did-eqt-opa-primer.md#securing-did-method-operations-with-opa
- https://datatracker.ietf.org/doc/html/draft-birkholz-scitt-architecture
- https://www.w3.org/2022/07/pressrelease-did-rec.html.en
- https://mastodon.social/@pdxjohnny/109804930974811967
- https://docs.microsoft.com/en-us/azure/confidential-ledger/architecture
- In search of more easy options to faciliate public/private hybrid chains of supply chain data. ideally OSS (SSI Service?).
- Similar work to address
- We first issue the verifiable credentials, with a reference to the rest of the hashlinks (or whatever the CBOR style thing is)
- Open Source Decentralized Identifiers and Verifiable Credentials Infrastructure and Tooling - verifiable-data/Hashlink.ts at c80cab9abe4db478add16b14837ba9a3afc3a70f
- Then we hashlink that, and that's the top level ref. effectivly the evolution of the zephyr.stripped shasum patchset
- zephyrproject-rtos/zephyr#51954
- Future
- As a follow on to the OpenSSF Metrics use case document and Living Threat Models are better than Dead Threat Models Rolling Alice: Volume 1: Coach Alice: Chapter 1: Down the Dependency Rabbit-Hole Again will cover how we identify and query provenance on dependencies where caching on data flow execution is assisted via quering public SCITT infrastructure and sourcing cached state from trustworthy parties.
- Leveraging our restoration of cached state from trustworthy parties and
LTM policy we can measure alignment of ML model's used within BOM components
so as to report conceptual alignment with entity strategic plans / principles.
- We care about data provenance. This provenance could be for example on inference derived from provenance from training data and model training env and config. This will allow us to ensure the prioritizer make decisions based on Sprit of the law / aka intent based policy derived from Trinity of Static Analysis, Dynamic Analysis, and Intent.
- Living Threat Model threats, mitigations, trust boundaries as initial data set for cross domain conceptual mapping of the the trinity to build pyramid of thought alignment to strategic principles.
- One of our strategic plans / principles might say
- "We must be able to trust the sources of all input data used for all model training was done from research studies with these ethical certifications"
- This allows us to write policies (Open Policy Agent to JSON to DID/VC/SCITT translation/application exploration still in progress) for the organizations we form and apply them as overlays to flows we execute where context appropriate. These overlaid flows define the trusted parties within that context as applicable to the active organizational policies as applicable to the top level system context.
- The policy associated with the principle that consumes the overlaid trust attestations we will implement and LTM auditor for which checks the SCITT provenance information associated with the operation implementations and the operation implementation network, input network, etc. within the orchestrator trust boundary.
- https://time.crystals.prophecy.chadig.com
- https://twitter.com/OR13b/status/1621907110572310528
- Actor
acquire
attachments
Link
toactivitypubextensions
threadcontent: "activitypubextensions"
threadinReplyTo: "$activitypubextensions_thread", content: "https://time.crystals.prophecy.chadig.com/bulk.1.0.0.schema.json"
thread- This becomes analogous to shared stream of consciousness uniform API for submitting across contexts (Manifests).
- CI/CD across projects with different orchestrators for downstream validation of the 2nd and 3rd party plugin ecosystem.
- This facilitates communication across pipelines across repos across PRs so we can use versioned learning to promote across trust boundaries (3rd party to 2nd party or support level 2 to 1)
- #1207
- #1315
- Alice helps us see risk over time, this is where we see Coach Alice, cartography used applied to dev branches, we grow closer to distributed compute with this, as iteration time is on dev branches rather than release or main
- This will probably be part of Alice and the Health of the Ecosystem
- CI/CD across projects with different orchestrators for downstream validation of the 2nd and 3rd party plugin ecosystem.
- This becomes analogous to shared stream of consciousness uniform API for submitting across contexts (Manifests).
- Ask him to reply to
@[email protected]
- Actor
- ActivityPub Actor watches for messages replying to certain threads
- Actor creates pull request to https://github.com/OR13/endor style repo
- Actor creates didme.me and gets VC SCITT receipt for associated
did:pwk:
(committed into Endor fork, he'd used git as database)- This could also be our content address of something in oras.land
- In the AI training data/human case we see the input data (meme) validated via SCITT
- We want to enable application of policy to data set ingestion, because this will happen in MLOps aka CI/CD
- Workstream: AI Ethics
- In the CI/CD use case, we see the input data (manifest referenced content, images, packages, metrics data output
FROM scratch
OpenSSF metrics use case) validated via SCITT. - Later we build up the threat modeling for the dynamic analysis portion of Alice which plays with input data as changes to repos and connects more of our Data, Analysis, Control for the software development process.
- Actor replies to Orie's reply with his receipt for his time crystals.
- Actor creates didme.me and gets VC SCITT receipt for associated
- For k8s style or OS DecentAlice style deployments (OSS scanning feeding OpenSSF metrics) we could run the graphed trust / event chain to a sidecar ActivityPub Actor / root of trust.
- https://twitter.com/OR13b/status/1621907110572310528
- RCFv1 Feedback
- https://mailarchive.ietf.org/arch/msg/scitt/dowMkmWhbi9Pkq6B5DhdXzip0so/
-
Maik: I'm curious to understand this more in detail. I think so far we've been thinking about the term federation as accepting SCITT claims/receipts from one transparency service in another transparency service, based on some policy. I think what you're describing is more about subscriptions/broadcasting/... Can you describe a little more on how you see this working in a SCITT transparency service? It might be that this is something that sits outside of it, but I'm not fully sure yet, so I wanted to understand it better first.
- Sketch response notes
- Policy as code, but sometimes it needs to execute because it's context aware, it's trying to decide for each piece of data that's pubsub'd for federation, does this SCITT instance want to be serving that trust assertion
- Open Policy Agent
- JSONLD
- Cypher
- Verfiable Credentials
-
SCITT claims/receipts from one transparency service in another transparency service, based on some policy
- Policy as compute contract, exec result determines entry of broadcast/subscribed syned SCITT instances (OSS projects, 2nd party, walled gardens)
- Can use this methodlolgy to broadcast those policies to topics via inReplyTo or later maybe service endpoint (as relay compute contact via did:oa)
- TDX self attest/DICE style layering where we wrap the receipt with the proposed insersion policy
- https://github.com/TrustedComputingGroup/DICE
- Something about the decentralized compute and the compute contract sandboxing to enable distributed trust propagation
- The AI travel in sandboxes. This is how trust (Alice) surfs the hypergraph. 🏄♀️ It's like ePBF but IPVM but as k8s admission controller but on scitt /inbox
- What's the sandbox? The policy here, what's the policy? The compute contract defining what the schema?+OPA?+cypher and what are the content addresses of the upstream contracts needed to fulfill the query
- On of the goals with the SCITT federation via ActivityPub is that it's a step towards the event stream being all JSONLD. Then audit and policy are effectively all done with definitions within DID referenced Verifiable Credentials. These encapsulate a receipt for a claim which who's insertion policy is a (or a context address of) policy as code aka compute contract. That contract statically defines or fulfils fetching or generating whatever data is needed to validate for insertion or federation and executes within a sandboxed environment. These policies can be overlayed with instance local additional policy as code. We can then read this event stream from anywhere or graft new trust chains off of it. GAUC is awesome it's just centralized from what I can tell, which is perfect for a performant view into a decentralized ecosystem. I think the two will work great together. We're all thinking in the same directions from what I can tell, just different goals in terms of data sovereignty, GUAC-GPT on the centralized side, Alice on the decentralized side.. The reason for the heavy focus on decentralization is that it for CI/CD we need to be able to spin dev and test chains of trust ad-hoc, for the AI side, we need to spin them for offline use cases tied to the users root of trust, or viewed as the user + their hardware roots of trust. Decentralized primitives allow us to never be forced to trust any authority other than what the deployment use case needs, scoping privilege to the threat model.
- This means that we can always do "one hop" analysis from this format set (DID+VC aka LDVC2)
- util: testing: manifest: shim: Initial commit intel/dffml#1273
- Alice status update
- This means that we can always do "one hop" analysis from this format set (DID+VC aka LDVC2)
- The AI travel in sandboxes. This is how trust (Alice) surfs the hypergraph. 🏄♀️ It's like ePBF but IPVM but as k8s admission controller but on scitt /inbox
- Something about it being recursive
- If your log inserted it and you want to federate that insertion to my log I want to know why you thought this was valid to insert. Tell me the scientific evidence Alice. What's the analysis report, why should that entry also get admission into my log. It's dynamic evaluation, so I can say based on my overlays okay I will take that receipt in my log based on that scientific method/process but due to this isntamces special overlays I won't take this other one. Maybe this is a recursion into the concept of a notary but required "self" notary required for that instances overlays applied to incoming receipt scientific processes / ipvm / data flow / open architecture.
- Decentralized trust propagation
- why should this be in your internal CA set sort of thing, what's your roots of trust and how to we justify adds / removes / modifications and notify that we think it's in your best interest for you to also trust the propagation of this trust assertion
- Decentralized dependency review - hence Open SSF Metrics use case
- Open Source maintiners general attidute: If you're going to tell my open source project what I should or should not do (use as a dependency) you better tell me why, or else the opne source repo / org SCITT we are propigating this to will deney based on it's execution of policy as code with it's overlays applied
- https://github.com/intel/dffml/blob/alice/docs/tutorials/rolling_alice/0001_coach_alice/0007_cartographer_extraordinaire.md
- Policy as compute contract, exec result determines entry of broadcast/subscribed syned SCITT instances (OSS projects, 2nd party, walled gardens)
- https://mailarchive.ietf.org/arch/msg/scitt/5SDINK63mr1BWX-BzbcsvalVLOA/
-
As part of this service offering example.com only allows artifacts to be added to the ledger by government customers who paid for the service. Example.com also requires a certain level of identity proofing and multi-factor authentication because it offers sensitive services to their government customers. Auditors require an even high level of identity proofing and authentication.
This policy of example.com is dictated by their business model and is outside the scope of what the IETF SCITT will define.
-
- https://mailarchive.ietf.org/arch/msg/scitt/cgz-9oif4SLMbdLyPn0P6-E8cIY/
-
This is interesting - many thanks Hannes. I notice our spec includes Merkle trees as the database structure - seems like an implementation detail, i.e. just a database. Can an implementer use, for example, an otherwise secured and RBAC'd record structure such as a file system or relational/hierarchical/sharded db, or is distributed ledger mandatory?
- df: overlay: Implement middleware/RBAC chains of ordered applications of overlays #1400
-
- There exist N instances of SCITT, which ones are tied to the current executing system context? Those determine reality, aka what is real, what to trust, for that system context
- Hallucinating large language models train/chain of thought is tied to context local SCITT. We graft in (accept new federated claims/receipts) from other system contexts we trust. We trust based on if we should propagate from that context to this context. Equilibrium as reality. Context local reality for the net. If they are isolated, they define their own reality.
- The truth will set you free. Equilibrium of record propagation.
- Everything is true, everything is permitted (insert policy*)
- Most of the time what's permitted is context dependent, overlays
- Time is context local, Earth is current most likely context, we have to ensure we factor in changes to this via abstraction of "when", ref lunar comms (LCRNS), beyond
- Vol 6: Where are your NTP gods now?
- https://esc.gsfc.nasa.gov/projects/LCRNS
- https://bigthink.com/hard-science/time-perception/
-
The idea of 'absolute time' is an illusion. Physics and subjective experience reveal why.
-
- Physics, also context local, gravity differs on different parts of Earth, and elsewhere. Make no assumptions, always policy as code, always recursive. Always show my why I should believe this message, message as the entity, not the messanger (instance federating from), focus on the truth in the message when propagating, not who propagates, critical thinking, threat modeling, deployment contexts
- Fork and exec over ActivityPub over DWN CLI
- https://github.com/soda480/wait-for-message-action
- TODO Fork and add ActivityPub support
- Usage of this in a
needs
->matrix/workflow_run/workflow_dispatch/ipvm
is below
- Closing the loop. Fork and exec via Distributed Compute, oras.land, VCS events ActivityPub over DWN (CLI) with DID VC SCITT with DID:KERI backed attested compute rooted identities. Enabling decentralized hardware roots of trust to facilitate decentralized asynchronous supply chains. Open source software development with end to end encrypted grafted chains communication only via attested channels for comms on vuln discovery with responsible disclosure and communication of remediation.
- This is also how we communicate
- Reference entity actions for this use case
- https://intel.github.io/dffml/main/examples/webhook/webhook.html#webhook-dataflow
- next tutorial: kcp -> k8s -> cf push -> webhook service -> dataflow to create activitypub event -> dwn-cli send -> webrtc -> dwn-cli recv ->
alice threats listen activitypub -stdin
->alice shouldi contribute
->alice please contribute
-> soft-serve/github repo pull request -> webhook service
- next tutorial: kcp -> k8s -> cf push -> webhook service -> dataflow to create activitypub event -> dwn-cli send -> webrtc -> dwn-cli recv ->
- https://www.youtube.com/watch?v=TMlC_iAK3Rg&list=PLtzAOVTpO2jYt71umwc-ze6OmwwCIMnLw&t=2064s
- https://www.youtube.com/watch?v=THKMfJpPt8I&list=PLtzAOVTpO2jYt71umwc-ze6OmwwCIMnLw&t=128s
- https://github.com/charmbracelet/soft-serve
- https://codeberg.org/forgejo/forgejo/issues/363
- Where is the best place to discuss federation of CI? Maybe in the spec repo? Shall I just throw up a pull request on that GitLab with the schema? We're interested in folks rebroadcasting their GitHub webhooks, etc. into the ActivityPub space so as to enable live at HEAD in poly repo envs (to help secure rolling releases).
- https://intel.github.io/dffml/main/examples/webhook/webhook.html#webhook-dataflow
- https://github.com/soda480/wait-for-message-action
Post RFCv3
- It's all the signal
- New data event
- inReplyTo context manifest, there was a new vex, there was a new registry image, there was a new something which will be interpreted in a context aware way based of reply set as label as input to running system context watching
- What is in the new data event?
- Where can I get the content (registry)
- How do I access it?
- New data event
- for gatekeeper policy options
- Related to reducing hallucinations in LLMs leveraging TEEs: https://queue.acm.org/detail.cfm?id=3623460
Upstream for S2C2F sections: https://github.com/ossf/s2c2f/blob/98803e0a558e6d8cef4d2770864ffd3cf7618c65/specification/framework.md#appendix-relation-to-scitt
The Supply Chain Integrity, Transparency, and Trust initiative, or SCITT, is a set of proposed industry standards for managing the compliance of goods and services across end-to-end supply chains. In the future, we expect teams to output "attestations of conformance" to the S2C2F requirements and store it in SCITT. The format of such attestations is to be determined.
S2C2F: Appendix: Mapping Secure Supply Chain Consumption Framework Requirements to Other Specifications
Goal: Create YAML file allowing users to map webhook event data to creation of data notarized by SCITT, statments and receipts created. Use YAML as basis for overall automatable format for alignment to S2C2F.
Requirement ID | Requirement Title | References |
---|---|---|
ING-1 | Use package managers trusted by your organization | CIS SSC SG : 3.1.5 OWASP SCVS: 1.2 CNCF SSC: Define and prioritize trusted package managers and repositories |
ING-2 | Use an OSS binary repository manager solution | OWASP SCVS: 4.1 CNCF SSC: Define and prioritize trusted package managers and repositories |
ING-3 | Have a Deny List capability to block known malicious OSS from being consumed | |
ING-4 | Mirror a copy of all OSS source code to an internal location | CNCF SSC: Build libraries based upon source code |
SCA-1 | Scan OSS for known vulnerabilities | SP800218 : RV.1.1 SP800161 : SA-10, SR-3, SR-4 CIS SSC SG : 1.5.5, 3.2.2 OWASP SCVS: 5.4 CNCF SSC: Verify third party artefacts and open source libraries, Scan software for vulnerabilities, Run software composition analysis on ingested software |
SCA-2 | Scan OSS for licenses | CIS SSC SG : 1.5.6, 3.2.3 OWASP SCVS: 5.12 CNCF SSC: Scan software for license implications |
SCA-3 | Scan OSS to determine if its end-of-life | SP800218 : PW.4.1 SP800161 : SA-4, SA-5, SA-8(3), SA-10(6), SR-3, SR-4 OWASP SCVS: 5.8 |
SCA-4 | Scan OSS for malware | clamav workflow or job |
SCA-5 | Perform proactive security review of OSS | SP800218 : PW.4.4 SP800161 : SA-4, SA-8, SA-9, SA-9(3), SR-3, SR-4, SR-4(3), SR-4(4) OWASP SCVS: 5.2, 5.3, |
INV-1 | Maintain an automated inventory of all OSS used in development | OWASP SCVS: 1.1, 1.3, 1.8, 5.11 CNCF SSC: Track dependencies between open source components |
INV-2 | Have an OSS Incident Response Plan | SP800218 : RV.2.2 SP800161 : SA-5, SA-8, SA-10, SA-11, SA-15(7) |
UPD-1 | Update vulnerable OSS manually | |
UPD-2 | Enable automated OSS updates | |
UPD-3 | Display OSS vulnerabilities as comments in Pull Requests (PRs) | |
AUD-1 | Verify the provenance of your OSS | CIS SSC SG : 3.2.4 OWASP SCVS: 1.10, 6.1 SLSA v1.0: Producing artifacts – Distribute provenance |
AUD-2 | Audit that developers are consuming OSS through the approved ingestion method | CIS SSC SG : 4.3.3 |
AUD-3 | Validate integrity of the OSS that you consume into your build | CIS SSC SG : 2.4.3 OWASP SCVS: 4.12 CNCF SSC: Verify third party artefacts and open source libraries |
AUD-4 | Validate SBOMs of OSS that you consume into your build | CNCF SSC: Require SBOM from third party supplier |
ENF-1 | Securely configure your package source files (i.e. nuget.config, .npmrc, pip.conf, pom.xml, etc.) | SP800218 : PO.5.2 CIS SSC SG : 2.4.2, 3.1.7, 4.3.4, 4.4.2 |
ENF-2 | Enforce usage of a curated OSS feed that enhances the trust of your OSS | SP800218 : PO.5.2 CIS SSC SG : 2.4.3, 3.1.1, 3.1.3 |
REB-1 | Rebuild the OSS in a trusted build environment, or validate that it is reproducibly built | CIS SSC SG : 2.4.4 |
REB-2 | Digitally sign the OSS you rebuild | SP800218 : PS.2.1 |
REB-3 | Generate SBOMs for OSS that you rebuild | SP800218 : PS.3.2 SP800161 : SA-8, SR-3, SR-4 CIS SSC SG : 2.4.5 OWASP SCVS: 1.4, 1.7 CNCF SSC: Generate an immutable SBOM of the code |
REB-4 | Digitally sign the SBOMs you produce | CIS SSC SG : 2.4.6 |
FIX-1 | Implement a change in the code to address a zero-day vulnerability, rebuild, deploy to your organization, and confidentially contribute the fix to the upstream maintainer |
Targets are new commits, branches, tags, and their CI/CD (status check) results
- Transforms GitHub webhook payloads into statements
- https://github.com/in-toto/attestation/blob/99d851228fe284c66b2cde353a6693c5eff69db1/spec/v1/statement.md
- https://github.com/in-toto/attestation/blob/99d851228fe284c66b2cde353a6693c5eff69db1/spec/predicates/test-result.md
- https://github.com/in-toto/attestation/blob/99d851228fe284c66b2cde353a6693c5eff69db1/spec/predicates/cyclonedx.md
- https://docs.github.com/en/rest/repos/contents?apiVersion=2022-11-28#download-a-repository-archive-tar
{
"_type": "https://in-toto.io/Statement/v1",
"subject": [
{
"name": "<NAME>",
"digest": {"<ALGORITHM>": "<HEX_VALUE>"}
},
...
],
"predicateType": "https://in-toto.io/attestation/test-result/v0.1",
"predicate": {
"result": "PASSED|WARNED|FAILED",
"configuration": ["<ResourceDescriptor>", ...],
"url": "<URL>",
"passedTests": ["<TEST_NAME>", ...],
"warnedTests": ["<TEST_NAME>", ...],
"failedTests": ["<TEST_NAME>", ...]
}
}
{
"_type": "https://in-toto.io/Statement/v1",
"subject": [
{
"name": "<NAME>",
"digest": {"<ALGORITHM>": "<HEX_VALUE>"}
},
...
],
"predicateType": "https://spdx.dev/Document/v2.3",
"predicate": {
"SPDXID" : "SPDXRef-DOCUMENT",
"spdxVersion" : "SPDX-2.3",
...
}
}