o1-preview
. I've used it for various tasks from writing documentation, specs, shell scripts, to code (in Go).The result? Well I can certainly say the model(s) are much better than they used to be, but maybe that isn't so much the models per se, but the sheer processing power at OpenAI's data centers? ๐ค
But here's the kicker though... If anyone ever for a moment ever think that these "AI" things are intelligent, or that the marketing and hype is ever remotely close to trying to convince of us this "AGI" (Artificial General Intelligence) or ASI (Artificial Super Intelligence), you are sorely mistaken.
Chat-GPT and basically and any other technology based on Generative-AI (Gen-AI), these pre-trained transformers that use adversarial neural networks and insanely multi-dimensional vector databases to model all sorts of things from human language, programming languages all the way to visual and audible art are (_wait for it_):
Incredibly stupid! ๐คฆโโ๏ธ
They are effectively quite useless for anything but:
- Reproducing patterns (_albieit badly_)
- Search and Retrieval (_in a way that "seems" to be natural_)
And that's about it.
Used as a tool, they're kind of okay, but I wouldn't use Chat-GPT or CoPilot. I'd stick with something more like Codeium if you want a bit of a fancier "auto complete". Otherwise, just forget about the whole thing honestly. It doesn't even really save you time.
o1-preview
. I've used it for various tasks from writing documentation, specs, shell scripts, to code (in Go).The result? Well I can certainly say the model(s) are much better than they used to be, but maybe that isn't so much the models per se, but the sheer processing power at OpenAI's data centers? ๐ค
But here's the kicker though... If anyone ever for a moment ever think that these "AI" things are intelligent, or that the marketing and hype is ever remotely close to trying to convince of us this "AGI" (Artificial General Intelligence) or ASI (Artificial Super Intelligence), you are sorely mistaken.
Chat-GPT and basically and any other technology based on Generative-AI (Gen-AI), these pre-trained transformers that use adversarial neural networks and insanely multi-dimensional vector databases to model all sorts of things from human language, programming languages all the way to visual and audible art are (_wait for it_):
Incredibly stupid! ๐คฆโโ๏ธ
They are effectively quite useless for anything but:
- Reproducing patterns (_albieit badly_)
- Search and Retrieval (_in a way that "seems" to be natural_)
And that's about it.
Used as a tool, they're kind of okay, but I wouldn't use Chat-GPT or CoPilot. I'd stick with something more like Codeium if you want a bit of a fancier "auto complete". Otherwise, just forget about the whole thing honestly. It doesn't even really save you time.
- Update the Twt Hash extension.
- Increase its truncation from 7 to 12
@xuu is right about quite a few things, and I'd love it if he wrote up the dynamic hash size proposal, but I'm inclined to just increase the length in the first place mostly because my own client
yarnd
doesn't even store the full hashes in the first place ๐คฆโโ๏ธ (I thinnk)
- Update the Twt Hash extension.
- Increase its truncation from 7 to 12
@xuu is right about quite a few things, and I'd love it if he wrote up the dynamic hash size proposal, but I'm inclined to just increase the length in the first place mostly because my own client
yarnd
doesn't even store the full hashes in the first place ๐คฆโโ๏ธ (I thinnk)
$ printf "%s\\t%s\\t%s" "https://example.com/twtxt.txt" "2024-09-29T13:30:00Z" "Hello World!" | sha256sum | awk '{ print $1 }' | xxd -r -p | base64 | head -c 12
UWVFdUXtvoLS
$ printf "%s\t%s\t%s" "https://example.com/twtxt.txt" "2024-09-29T13:30:00Z" "Hello World!" | sha256sum | awk '{ print $1 }' | xxd -r -p | base64 | head -c 12
UWVFdUXtvoLS
$ printf "%s\t%s\t%s" "https://example.com/twtxt.txt" "2024-09-29T13:30:00Z" "Hello World!" | sha256sum | awk '{ print $1 }' | xxd -r -p | base64 | head -c 12
UWVFdUXtvoLS
sha256sum
vs. b2sum
. Neither is more complicated than the other.
sha256sum
vs. b2sum
. Neither is more complicated than the other.
=> https://gist.mills.io/prologic/194993e7db04498fa0e8d00a528f7be6
e.g: (_turns out @xuu is right about Blak2b being easy/simple too!_):
$ printf "%s\\t%s\\t%s" "https://example.com/twtxt.txt" "2024-09-29T13:30:00Z" "Hello World!" | b2sum -l 32 -t | awk '{ print $1 }'
7b8b79dd
=
=> https://gist.mills.io/prologic/194993e7db04498fa0e8d00a528f7be6
e.g: (_turns out @xuu is right about Blak2b being easy/simple too!_):
$ printf "%s\t%s\t%s" "https://example.com/twtxt.txt" "2024-09-29T13:30:00Z" "Hello World!" | b2sum -l 32 -t | awk '{ print $1 }'
7b8b79dd
=
=> https://gist.mills.io/prologic/194993e7db04498fa0e8d00a528f7be6
e.g: (_turns out @xuu is right about Blak2b being easy/simple too!_):
$ printf "%s\t%s\t%s" "https://example.com/twtxt.txt" "2024-09-29T13:30:00Z" "Hello World!" | b2sum -l 32 -t | awk '{ print $1 }'
7b8b79dd
=
> iirc in twtxt v2 it starts prohibited
This is not true. There are no issues supporting fetching feeds via Gemini/Gopher. This is totally fine. What will likely happen is "recommendations" and "drawbacks of using Gemini/Gopher"
> iirc in twtxt v2 it starts prohibited
This is not true. There are no issues supporting fetching feeds via Gemini/Gopher. This is totally fine. What will likely happen is "recommendations" and "drawbacks of using Gemini/Gopher"

