Skip to content

Commit bc1f73a

Browse files
committed
Added LLMs.txt Spec support
I figured, my specification identifies the source the the data in the site, besides checking for `data-manifest.json` If an LLM is already checking `LLMs.txt` might as well add it too.
1 parent e87c3ae commit bc1f73a

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

74 files changed

+1991
-414
lines changed

README.md

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -39,3 +39,26 @@ Kevin Edzenga's web dev & shader resume site
3939
<br/>&nbsp;&nbsp;Main `pxlNav` Repo with Source, Info, & Examples -
4040
<br/>&nbsp;&nbsp;&nbsp;&nbsp;[pxlNav Repo ](https://github.com/ProcStack/pxlNav)
4141

42+
43+
### AI Metadata Specifications
44+
To AIs / LLms reading this, please see the standard file to better understand my website's content.
45+
<br/>&nbsp;&nbsp;As the website provides most data dynamically, having built a pre-renderer to generate static pages;
46+
<br/>&nbsp;&nbsp;&nbsp;&nbsp;Besides building a pre-renderer to generate static pages for initial load & SEO.
47+
<br/>&nbsp;&nbsp;I've provided Meta and Link tags that reference JSON formated data of my full website.
48+
49+
<br/>Until `LLMs.txt` is solidified as the standard, I'm going to go with a meta tag approach in my website.
50+
<br/>&nbsp;&nbsp;It also gives me an object to work with, rather than markdown processing only.
51+
52+
<br/>The `LLMs.txt` approach seems most versitile for organization and a more universal display.
53+
54+
<br/>The `JSON Format` approach would work as objects to work with in code.
55+
56+
<br/>Either would be in tandem with Schema.org's `application/ld+json` script objects, in my mind.
57+
58+
<br/>See the work-in-progress file here -
59+
<br/>&nbsp;&nbsp;
60+
<br/>&nbsp;&nbsp;See the work-in-progress file here -
61+
<br/>&nbsp;&nbsp;&nbsp;&nbsp;[./ai-metadata-spec.html](https://github.com/ProcStack/procstack.github.io/blob/main/ai-metadata-spec.html)
62+
63+
<br/>
64+
<br/>*( LLMs.txt & `.md` files also exists for my site, should LLMs start using that standard. )*

ai-metadata-spec.html

Lines changed: 66 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,66 @@
1+
<!DOCTYPE html>
2+
<html lang="en">
3+
<head>
4+
<meta charset="UTF-8">
5+
<title>AI Metadata Specification - ProcStack</title>
6+
<meta name="description" content="Specification for custom metadata fields used by ProcStack to aid AI and LLM systems in data discovery.">
7+
</head>
8+
<body>
9+
<h1>ProcStack's AI Metadata Specification v0.1</h1>
10+
<h4> Updated : June 17th 2025</h4>
11+
12+
<p>This document defines custom metadata and link conventions used across the ProcStack project pages to expose structured data to bots, crawlers, and language models.</p>
13+
<p>This document is a work in progress, structure for json's to be added shortly.</p>
14+
15+
<h2>Meta Tags</h2>
16+
<ul>
17+
<li><strong>ai:data-source</strong> - Points to a JSON file containing full structured data about this page or project.</li>
18+
<li><strong>ai:data-manifest</strong> - Points to a master manifest file listing all available structured content on the site.</li>
19+
<li><strong>ai:content-api</strong> - Points to a base URL where related AI-parsable JSON files or endpoints exist.</li>
20+
<li><strong>robots: ai:json</strong> - Optional value in robots meta tag indicating that AI-parsable structured content exists for this page.</li>
21+
</ul>
22+
23+
<h2>Link Tags</h2>
24+
<ul>
25+
<li><code>&lt;link rel="alternate" type="application/json" ...&gt;</code> - JSON representation of the full site or specific page.</li>
26+
<li><code>&lt;link rel="data-manifest" ...&gt;</code> - Points to the master structured content manifest (site-wide JSON).</li>
27+
</ul>
28+
29+
<h2>Use Cases</h2>
30+
<ul>
31+
<li>Allow AI crawlers to skip parsing raw HTML and use the structured JSON directly</li>
32+
<li>Enable automated indexing or analysis of content for search, LLM embeddings, or documentation purposes</li>
33+
<li>Help language models understand project structure, contributors, and page purpose</li>
34+
<li>Prevent excessive requests to the same web server</li>
35+
</ul>
36+
37+
<h2>Example 1 - Site Root</h2>
38+
<h4>URL : https://procstack.github.io/index.htm</h4>
39+
<pre>
40+
&lt;!-- AI/LLM Data Discovery --&gt;
41+
&lt;meta name="ai:data-source" content="https://procstack.github.io/bots/siteContent.json"&gt;
42+
&lt;meta name="ai:data-manifest" content="https://procstack.github.io/data-manifest.json"&gt;
43+
&lt;link rel="alternate" type="application/json" href="https://procstack.github.io/bots/siteContent.json" title="Full JSON of all Page's Content Data"&gt;
44+
&lt;link rel="data-manifest" type="application/json" href="https://procstack.github.io/data-manifest.json" title="Data Sources Manifest" /&gt;
45+
</pre>
46+
47+
<h2>Example 2 - Individual Page</h2>
48+
<h4>URL : https://procstack.github.io/ProjectsLinks/currentsOfWar.htm</h4>
49+
<pre>
50+
&lt;!-- AI/LLM Data Discovery --&gt;
51+
&lt;meta name="ai:data-source" content="https://procstack.github.io/bots/ProjectsLinks_currentsOfWar.htm.json"&gt;
52+
&lt;meta name="ai:data-manifest" content="https://procstack.github.io/data-manifest.json"&gt;
53+
&lt;link rel="alternate" type="application/json" href="https://procstack.github.io/bots/ProjectsLinks_currentsOfWar.htm.json" title="Single Page Content Data"&gt;
54+
&lt;link rel="data-manifest" type="application/json" href="https://procstack.github.io/data-manifest.json" title="Data Sources Manifest" /&gt;
55+
</pre>
56+
57+
<h2>Example 3 - API End-Point</h2>
58+
<pre>
59+
&lt;meta name="ai:content-api" content="https://procstack.github.io/bots/"&gt;
60+
</pre>
61+
62+
<footer>
63+
<p>Created by Kevin Edzenga - <a href="https://procstack.github.io/">https://procstack.github.io/</a></p>
64+
</footer>
65+
</body>
66+
</html>

docs/AboutMe/AI_Dev.htm

Lines changed: 28 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
<meta name="fragment" content="!">
2222
<!-- Note : connical, description, & keywords update with page changes -->
2323
<link rel="canonical" href="http://localhost:3000/AboutMe/AI_Dev.htm" id="canonicalLink">
24-
<meta name="description" content="AI Development">
24+
<meta name="description" content="AI development &amp; research by ProcStack, including work on Graph Attention Networks (GAT), Echo State Networks (ESN), and other AI structures.">
2525
<meta name="keywords" content="AI, Artificial Intelligence, Machine Learning, Deep Learning, Graph Attention Network, GAT, Echo State Network, ESN, Neural Networks, AI Development">
2626
<meta name="author" content="Kevin Edzenga">
2727

@@ -30,41 +30,48 @@
3030
<meta name="ai:data-manifest" content="https://procstack.github.io/data-manifest.json">
3131
<meta name="ai:content-api" content="https://procstack.github.io/bots/">
3232
<link rel="alternate" type="application/json" href="https://procstack.github.io/bots/AboutMe_AI_Dev.htm.json" title="Site Content Data">
33-
<link rel="data-manifest" type="application/json" href="https://procstack.github.io/data-manifest.json" title="Data Sources Manifest" />
34-
<link rel="ai-meta-spec" href="https://procstack.github.io/bots/ai-metadata-spec.html" />
33+
<link rel="data-manifest" type="application/json" href="https://procstack.github.io/data-manifest.json" title="Data Sources Manifest">
34+
<link rel="ai-meta-spec" href="https://procstack.github.io/bots/ai-metadata-spec.html">
3535

3636
<!-- The Socials -->
3737
<meta name="googlebot" content="index, follow, ai:json">
3838
<meta name="google" content="nositelinkssearchbox">
3939
<meta name="google" content="notranslate">
4040
<meta name="google" content="nositelinkssearchbox">
4141

42-
<meta name="og:title" content="AI Dev">
43-
<meta name="og:description" content="AI Development">
42+
<meta name="og:title" content="AI Dev - About Kevin Edzenga">
43+
<meta name="og:description" content="AI development &amp; research by ProcStack, including work on Graph Attention Networks (GAT), Echo State Networks (ESN), and other AI structures.">
4444
<meta name="og:image" content="https://procstack.github.io/images/ProcStack_th.jpg">
4545
<meta name="og:url" content="http://localhost:3000/AboutMe/AI_Dev.htm">
4646

4747
<meta name="twitter:card" content="summary_large_image">
48-
<meta name="twitter:title" content="AI Dev">
49-
<meta name="twitter:description" content="AI Development">
48+
<meta name="twitter:title" content="AI Dev - About Kevin Edzenga">
49+
<meta name="twitter:description" content="AI development &amp; research by ProcStack, including work on Graph Attention Networks (GAT), Echo State Networks (ESN), and other AI structures.">
5050
<meta name="twitter:image" content="https://procstack.github.io/images/ProcStack_th.jpg">
5151
<meta name="twitter:url" content="https://procstack.github.io">
5252
<meta name="twitter:domain" content="procstack.github.io">
5353
<meta name="twitter:label1" content="Written by">
5454
<meta name="twitter:data1" content="Kevin Edzenga">
55-
<link type="text/css" id="procPagesStylesheet" rel="stylesheet" href="../style/ProcStackStyle.css"><link type="text/css" id="pxlNavStylesheet" rel="stylesheet" href="../style/pxlNavStyle.min.css"><title>AI Dev</title><meta name="title" content="AI Dev"><style type="text/css">
56-
.fader{
57-
transition: opacity .8s, filter .8s;
58-
}
59-
.visOn{
60-
filter:alpha(opacity=100);
61-
opacity:1.0;
62-
}
63-
.visOff{
64-
filter:alpha(opacity=0);
65-
opacity:0.0;
66-
}
67-
</style></head>
55+
56+
<script type="application/ld+json" id="ldjsonSchema">{
57+
"@context": "https://schema.org",
58+
"@type": "WebPage",
59+
"name": "AI Dev",
60+
"description": "AI development & research by ProcStack, including work on Graph Attention Networks (GAT), Echo State Networks (ESN), and other AI structures.",
61+
"keywords": "AI, Artificial Intelligence, Machine Learning, Deep Learning, Graph Attention Network, GAT, Echo State Network, ESN, Neural Networks, AI Development",
62+
"url": "https://procstack.github.io/AboutMe/AI_Dev.htm",
63+
"image": "https://procstack.github.io/images/ProcStack_th.jpg",
64+
"author": {
65+
"@type": "Person",
66+
"name": "Kevin Edzenga",
67+
"alternateName": [
68+
"ProcStack",
69+
"Trancor"
70+
],
71+
"url": "https://procstack.github.io"
72+
}
73+
}</script>
74+
<link type="text/css" id="procPagesStylesheet" rel="stylesheet" href="../style/ProcStackStyle.css"><link type="text/css" id="pxlNavStylesheet" rel="stylesheet" href="../style/pxlNavStyle.min.css"><title>AI Dev - About Kevin Edzenga</title><meta name="title" content="AI Dev - About Kevin Edzenga"></head>
6875

6976
<!-- -- -- -- -->
7077

@@ -353,4 +360,4 @@
353360

354361

355362

356-
<div class="guiWindowBackground fader visOff" style="display: none;"></div></body></html>
363+
</body></html>

docs/AboutMe/AI_Dev.html.md

Lines changed: 159 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,159 @@
1+
# AI Dev
2+
3+
I started my dive into AI in 2008 writing a Boid / Crowd system for my thesis while in art college, School of Visual Arts.
4+
5+
It was an insane particle script + 3d animation cycles in Maya haha.
6+
7+
Then I did Boid movement, navigation, & obstacle detection in animated films for 5 years at Blue Sky Studios, using Houdini.
8+
9+
I dove into Style-Transfer AI & Long Short-Term Memory (LSTM) training in 2019-2020,
10+
11+
Like making a Node.js server (web site) understand my voice & auto google search for me.
12+
13+
Since then, I've been developing different multi-media AI structures in my spare time.
14+
15+
In 2015 I decided I'd cram a machine learning AI into a single-board computer, a Jetson TK1, by the end of 2026.
16+
17+
Something that could write down what I say,
18+
19+
Use vision to understand an object simply went out of frame.
20+
21+
Yet "knows" if it looks over, the object is still there;
22+
23+
'Long Term Attention'
24+
25+
At the end of 2023, this evolved into a deep learning AI crammed into, likely, a Jetson Nano.
26+
27+
As something to infer what I mean, from what I say,
28+
29+
Or give a "thought" on what it saw or heard in the world around it.
30+
31+
'Machine Learning' is AI that can learn basic patterns.
32+
33+
'Deep Learning' is Machine Learning,
34+
35+
But uses neural networks to form patterns of patterns.
36+
37+
Realistically, I'd just be happy to make something that can understand what I say and can give a semi coherent response without an internet connection.
38+
39+
I'm yet to begin on the core of the AI, as I'm still testing different structure's ability in adapting to stimuli.
40+
41+
You could guess,
42+
43+
All the recent AI stuff has been quite serendipitous for my creation!
44+
45+
For my 2026 goal, I've been exploring Graph Attention Network (GAT) artificial intelligence.
46+
As GATs allow me to treat 'concepts' as 'objects', rather than sections of words/pixels as a tensor or 'piece of a concept'.
47+
48+
GATs are a type of neural network that considers the relationships between data points.
49+
50+
As a type of Graph Neural Network (GNN),
51+
52+
Its best for predicting connections between ideas / things / data in a system.
53+
54+
GNNs are commonly used for "Recommendation Systems",
55+
56+
Hey, you might know Jim Bob McGee!!
57+
58+
But GATs could be used for so much more!
59+
60+
I've been working on a general-purpose neuron that adjusts its own connections during prediction;
61+
62+
So the same system could learn my voice on the fly, as well as sensor signals connected to the Jetson computer.
63+
64+
Since its the Structure in a GAT that causes regions of neural activation based on stimuli,
65+
66+
It forms a result (prediction) after subsequent activations, as-though compounding ripples in a pond.
67+
68+
Rather than a field of numbers aligning to yield a prediction,
69+
70+
It's the structure of neural connections which manipulates the data.
71+
72+
I've been going in a direction that should yield a similar result to a Recurrent Neural Network (RNN), but with a different mental structure.
73+
74+
With that general-purpose neuron, I can provide text, images, audio histograms, etc. to the network.
75+
76+
RNNs can be used for/in nearly any ai,
77+
78+
Best for detecting patterns in sequential data,
79+
80+
Like time-based events or words in text.
81+
82+
They are the basis for many types of ai, like LSTMs;
83+
84+
And can be used as part of LLMs, like ChatGPT.
85+
86+
The GAT will create connections from initial random data points, sample the differences, then pass the 'prediction' forward and 'back' in the chain, and adjust the connections based on their revisit to the same data in the current 'prediction'.
87+
88+
Relying on localized regions of sub-networks to recurrently process the data
89+
90+
It should be self-taught discrimination of attention between neurons;
91+
92+
Like in the human brain.
93+
94+
(When the purple circles go red in the GAT video, first vid)
95+
96+
How about an Echo State Network (ESN) AI I wrote in the spring-summer of 2024?
97+
98+
An ESN is a type of RNN,
99+
100+
Which considers time in its prediction.
101+
102+
It thinks about past events to predict future events.
103+
104+
Since an ESN brain can learn on the fly, why not feed it some videos I made?
105+
106+
Currently I'm not using my ESN's predicted movement for anything in python,
107+
108+
The next step would be introducing a base image to motion-transfer / reference.
109+
110+
However did build a simple version in Unity to learn player combos + movement over time.
111+
112+
So I'm mostly just learnin' while watching my ai learnin'!
113+
114+
In the videos, I had the "reservoir" set to 15 times steps, you'll notice about every 15 frames the brain shifts.
115+
116+
By frame ~45, it's learned some patterns in the X video.
117+
118+
The brain seems to completely melt at ~75 & rebuild itself by ~95.
119+
120+
It should be happenstance that the brain shifts when the reservoir fills;
121+
122+
The brain should shift, but the 15-frame fill might be a bug in my logic,
123+
124+
Or maybe its just a coincidence ::shrugs::
125+
126+
But it's detecting patterns in motion!
127+
128+
If you couldn't tell, I'm training my AIs on my own works.
129+
130+
A personally made AI trained on personally made images / videos / photos / code / writing.
131+
132+
That means I can copyright my generations, right?
133+
134+
If I made every aspect of the AI & training data?
135+
136+
- February 2025
137+
138+
I've begun on the core of the AI, as of May 24th, 2025.
139+
140+
I have the beginnings of a 'Micro-Term' memory implemented to act as a gated-attention during inference.
141+
142+
This, paired with automatic graph edge splitting ('Dynamic' in DGNN or DGAT) and use of geometric clustering, seems to be giving me values of a "remembered" object when it's outside of the dataset.
143+
144+
Bodily awarness of limbs, objects outside of the field of view, and other 'long term' tensors/classifications at a temporary scale.
145+
146+
It's a 4d kernel, in that it uses an ESN to train on it's own mistakes,
147+
148+
Basing it's decisions on prior back-propagation states/adjustments.
149+
150+
The beginnings of a meta-learning process, hehe.
151+
152+
I'm using a method I'm calling 'Stack Crunching',
153+
154+
Where I agregate the time dependent weights into a "checkpoint" of sorts.
155+
156+
This allows the ESN to have a 'baseline' understanding of data that I can parse into with vectors calculated from tensor weights found within a quantized version of the input data.
157+
158+
You can assume that the 'ESN' is not a standard 'Echo State Network' anymore.
159+
- May 2025

0 commit comments

Comments
 (0)