I was up till midnight working on it, so I’ll keep this brief:
- The agent processing algorithm is much faster now.
- I cleaned up the code, getting rid of a bunch of clutter and making the logic more elegant.
- Sequence recognition now works to any depth. For instance: a word is a sequence of letters, a phrase is a sequence of words, a sentence is a sequence of phrases. Recognizing deeply “nested” sequences like this turns out to be a lot more complicated than you might expect. My previous algorithm was supposed to handle it, but it was convoluted and buggy. The new version is cleaner and (so far) works 100%.
- The AI can recognize certain parts of speech (nouns, adjectives, verbs) and construct grammatical relationships between them (e.g. this adjective modifies this noun). It can then use other agents to map the grammar structure (adjective “black” modifies noun “dog”) to a semantic structure (entity “dog” has property “black”). So far this is still in its early stages, but the foundation for building it up is solid.
- Text-to-speech is fully implemented by calling a free online text-to-speech API. Before, it could type; now, it can talk. Click here for a sample of what its voice sounds like. That’s not a static mp3 file. It’s generated dynamically based on the text you give it.
- I gave it the ability to load websites on command. So I can type “google” and it can bring up the Google web page. Obviously that’s nothing special in itself, but it’s significant because website-loading is now fully integrated with its other capabilities. For the AI, loading a site is like flexing a muscle.
Gotta run. If you have any questions, just ask!