▲ | gabriel666smith 3 days ago | |||||||
> Where are you going with this? I don't know! Adding code now. Is this more clear, for at least the initial 'word set' generation? I can add to repo if so: Concept: The system predicts and generates words based on Fibonacci distances - instead of looking at the next word or previous word, it looks at words that are 2, 3, 5, 8, 13, 21, etc. positions away (following the Fibonacci sequence). Key Components 1. Training Phase Takes a text file and extracts all words Builds two prediction models: Forward model: "If I see word X at position N, what word appears at position N+2, N+3, N+5, N+8, etc.?" Backward model: "If I see word X at position N, what word appeared at position N-2, N-3, N-5, N-8, etc.?" 2. Generation Phase Starts with seed words (user input) For each seed word, predicts what should come before and after using Fibonacci distances Uses bidirectional validation: a word is only chosen if it's probable in BOTH forward and backward directions This attempts to create a more coherent, contextually consistent text. Then runs multiple passes where generated words become new starting points for further generation, creating richer, more developed text. The words with strongest association values = the final generation set of available words. | ||||||||
▲ | jstanley 3 days ago | parent [-] | |||||||
When you say you build "prediction models" - what exactly is that? Is it just a list of possibilities for each word at each position? So for example forward[the][2] is a list of words that can come 2 places after "the"? Either with duplicates for more likely words, or with probabilities that you can sample, like a Markov model. Or is the "prediction model" some sort of neural network, or something else? When you say a word is only chosen if it's probable in both the forward and backward direction, what does that mean? I still can't see any code in your repo. | ||||||||
|