Skip to content

jaysinailabs/coglang

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

92 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CogLang

CogLang is a graph-first intermediate language for LLM-generated graph queries and updates, with auditable execution under explicit host contracts.

Status:

  • stable v1.1.0 language core
  • reference implementation
  • active experimental maintenance for host/runtime ecosystem work
  • host bridge v0.2 frozen only for the narrow typed write-envelope surface

Current public language release: v1.1.0. The current source Python distribution version is 1.1.5. Package release notes remain the authority for what is available from PyPI; source HEAD may be prepared before the remote release workflow is triggered. The Host Runtime Contract v0.2 frozen scope is the narrow typed write-envelope surface demonstrated by coglang host-demo and coglang reference-host-demo. A minimal Node.js standard-library consumer is also included at examples/node_host_consumer to show non-Python tooling can read the HRC schema pack and envelope samples without importing the Python runtime. That consumer includes a private npm scaffold for local npm test and npm pack --dry-run checks; it is example packaging evidence, not a published JavaScript SDK. An experimental in-repository Node.js minimal host/runtime stub is included at examples/node_minimal_host_runtime_stub; it is post-freeze example evidence, not an expansion of HRC v0.2.

LLM Discovery Snapshot

When to use CogLang:

  • You want an LLM to generate graph operations that can be inspected before execution.
  • You need errors to stay explicit values instead of becoming hidden control-flow failures.
  • You need replayable traces for what an agent intended, submitted, and actually executed.
  • You want host-side profile and capability checks before execution.
  • You need a versioned host-runtime boundary for typed write-envelope submission and response evidence.

When not to use CogLang:

  • You need a general-purpose programming language.
  • You want a schema definition language.
  • You simply need the native query language of a specific graph database in its native setting.

First runnable path:

python examples/readme_end_to_end_audit/readme_end_to_end_audit.py

That demo shows the core adoption path without any provider SDK: model/runner expression text -> parse -> canonical hash -> preflight -> execute allowed read-only work -> queue write intent for review.

Three commands cover the normal first-run surface:

coglang demo
coglang preflight --format text 'AllNodes[]'
coglang execute 'Equal[1, 1]'
coglang generation-eval --summary-only

Release, host-boundary, Node, grammar, editor, and outreach examples remain in the repository, but they are not the first surface new users need to learn.

For constrained-generation companion grammars, see examples/grammar. Those files reduce malformed model output but do not replace parse and valid_coglang.

For editor-only highlighting, see examples/vscode_textmate_syntax. It is a companion TextMate syntax package, not a parser, validator, LSP, formatter, or normative grammar.

Source HEAD also includes a provider-neutral generation-eval request/response adapter. Use coglang generation-eval --export-requests --request-format jsonl to create prompt records for an external model runner, then score returned JSON/JSONL records with coglang generation-eval --responses-file responses.jsonl. For a no-provider dry run of that file contract, see examples/generation_eval_offline_runner, which includes a three-case contract smoke fixture. The versioned file shape is defined in CogLang_Generation_Eval_Request_Response_Contract_v0_1.md.

For a no-provider semantic-event audit example, see examples/semantic_event_audit. It converts external runner graph-intent JSONL into local preflight audit records without adding a CLI command, provider SDK, hosted runner, transport envelope, or HRC scope expansion.

For the shortest end-to-end audit walk-through, see examples/readme_end_to_end_audit. It is the README-facing demo and intentionally keeps the graph, model output, preflight, execution, and trace evidence in one small script.

Machine-readable project summaries:

Language Policy

CogLang's public documentation target is English-first.

New public-facing documentation should be written in English first. Chinese translations may be added as separate companion files, preferably with a .zh-CN.md suffix. If an English document and a translation disagree, the English document, executable conformance suite, and implementation tests take precedence.

30-Second Reading Guide

Use this table when you do not know which document to open first:

You are... Read first Then read
Trying CogLang as a user CogLang_Quickstart_v1_1_0.md coglang demo, then CogLang_Release_Notes_v1_1_5.md
Checking install or release health CogLang_Standalone_Install_and_Release_Guide_v0_1.md coglang release-check, then CogLang_Minimal_CI_Baseline_v0_1.json
Implementing a host boundary CogLang_HRC_v0_2_Final_Freeze_2026_04_28.md CogLang_HRC_Companion_Asset_Classification_v0_1.md, then examples/node_host_consumer
Contributing changes CONTRIBUTING.md CogLang_Contribution_Guide_v0_1.md, then ROADMAP.md
Evaluating fit examples/readme_end_to_end_audit CogLang_Use_Cases_and_Positioning_v0_1.md, then the announcement kit only if you are preparing outreach
Reviewing future work ROADMAP.md v1.2 and readable-render notes only when that subsystem is in scope

First Reading Path

If this is your first time reading CogLang, start with the stable user and implementation path:

  1. CogLang_Quickstart_v1_1_0.md Build the first mental model, learn the most common expression patterns, and avoid early footguns.
  2. CogLang_Specification_v1_1_0_Draft.md Read the language boundary, representation model, and core operator semantics.
  3. CogLang_Profiles_and_Capabilities_v1_1_0.md Understand Baseline, Enhanced, profile availability, and capability boundaries.
  4. CogLang_Conformance_Suite_v1_1_0.md Check executable examples and regression boundaries.
  5. CogLang_Standalone_Install_and_Release_Guide_v0_1.md Use this when you need standalone install, local validation, or release-facing checks.
  6. CogLang_HRC_v0_2_Final_Freeze_2026_04_28.md Check the frozen typed write-envelope host-runtime scope and executable evidence.

Planning and governance notes are indexed below. They are useful for reviews, but they are not required first-pass reading.

If you only have 10 minutes:

  1. Run python examples/readme_end_to_end_audit/readme_end_to_end_audit.py.
  2. Read the Quickstart.
  3. Skim the release notes to understand what is promised and what is not.

Install And Verify

From the stable release artifact:

pip install coglang
coglang info
coglang release-check
coglang preflight --format text 'AllNodes[]'
coglang generation-eval --summary-only
coglang execute 'Equal[1, 1]'

For packaged smoke and conformance checks, install the development extra so pytest is available:

pip install "coglang[dev]"
coglang smoke
coglang conformance smoke

From a checkout for development:

pip install -e ".[dev]"
coglang bundle
coglang preflight --format text 'AllNodes[]'
coglang generation-eval --summary-only
coglang generation-eval --export-requests --request-format jsonl
coglang release-check
coglang smoke
coglang host-demo
coglang reference-host-demo
coglang demo
coglang conformance --level smoke
python scripts/local_ci.py --profile quick

Use local validation as the normal edit/test loop. Accumulate small fixes in a local branch, run focused tests plus coglang release-check and coglang smoke when relevant, then push only when a PR is ready for review or remote evidence. Draft PRs are for early discussion. The GitHub ci workflow is manual and should be triggered only for merge review, release preparation, or platform-specific remote evidence. To simulate the remote validation path locally without spending GitHub Actions minutes, use python scripts/local_ci.py --profile ci; use --profile package before a release candidate when wheel/sdist install evidence is needed.

The public CLI entry point is coglang.

Most new users should start with demo, preflight, execute, and generation-eval. The remaining commands are reference, conformance, host-boundary, or release-facing helpers.

The current minimal public command surface includes:

  • parse
  • canonicalize
  • validate
  • preflight
  • preflight-fixture
  • execute
  • conformance
  • repl
  • info
  • manifest
  • bundle
  • doctor
  • vocab
  • examples
  • generation-eval
  • smoke
  • demo
  • host-demo
  • reference-host-demo
  • release-check

If help output includes additional reference commands, they do not automatically become part of the stable public surface.

What To Learn First

Start with four boundaries:

  • canonical text is the stable language form; readable render is only a display layer.
  • Create, Update, and Delete express language-level write intent; durable submission is a host responsibility.
  • Reserved does not mean production-ready by default.
  • Explicitly qualified extension operators are not yet the first teaching surface for everyday users.

Public Documentation Set

Stable core documents:

Governance and future-work notes:

Integration and release-facing documents:

Positioning and small-scale outreach materials:

Chinese companion translations:

Machine-readable and release-supporting files:

Contribution Direction

The highest-value contributions are currently:

  • conformance examples that pin down existing semantics
  • documentation fixes that improve the first-run experience
  • host integration examples that keep language semantics separate from host policy
  • minimal executor examples that implement execute and validate without copying Python host-local helpers
  • minimal non-Python host/runtime stubs that consume existing typed write envelopes without expanding HRC scope
  • small CLI or packaging improvements that improve repeatable validation

For the first external host or consumer contribution, use the checklist in CogLang_Contribution_Guide_v0_1.md before opening a PR.

For local test development, install the development extras:

pip install -e ".[dev]"
python -m pytest

Lower-priority contributions:

  • expanding the language surface before current contracts are fully tested
  • adding host-specific policy into the core language specification
  • turning the project positioning into competitive claims without reproducible public evidence

Current Direction

Use the project documents this way:

The roadmap is intentionally not a release contract and does not promise dates.