RPG Word Embeddings – What is Most Similar? – 01

I made a gloVe embedding model based on my game book collection – 7000 odd, of which 6000 or so managed to make it through a first pass pdf extraction pipeline

This framework is quite good https://github.com/NRCan/geoscience_language_models/tree/main/project_tools

https://github.com/NRCan/geoscience_language_models/tree/main/project_tools and parallelises, which is important for big books

The C version of gloVe is superior:

https://github.com/stanfordnlp/GloVe

With some work you can get a python version going, but I wouldn’t recommend for large numbers.

e.g. https://pypi.org/project/glove-py

and associated hacks..

The Notebook associated with this is here: https://github.com/bluetyson/RPG-gloVe-Model

These days microsoft probably won’t let you see something that big online, so will make a series of post excerpts.

Freedom In the Deep

https://www.australiangeographic.com.au/wp-content/uploads/2023/02/DSC_9482-SharpenAI-Standard-WIDE-1800×1702.jpeg

This would be good to turn a fantasy adventure version of into a megaladon.

https://www.textpublishing.com.au/books/big-meg-the-story-of-the-largest-and-most-mysterious-predator-that-ever-lived
Picture of woman freedriving with a whale shark

Zork Lives!

https://arstechnica.com/gaming/2023/11/infocoms-ingenious-code-porting-tools-for-zork-and-other-games-have-been-found/

Or will soon, anyway, if tools to do so have been found. I have been reading the 50 Years of Text Games Kickstarter book.

%d bloggers like this: