Is there an issue? Send a MessageReason:
None
Changed line(s) 11 (click to see context) from:
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (any finetuning is done offline at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
to:
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (any finetuning is done offline at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.time.
----
----
Is there an issue? Send a MessageReason:
General → Generative Pretrained Transformer
Changed line(s) 3,4 (click to see context) from:
Its original version was powered by the XL-sized (1.5B internal parameters) '''General Pre-trained Transformer''' revision 2 (GPT-2), a neural network text predictor expensively ($ 150,000 USD) trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walton (creator of [=AID2=] and a previous incarnation that used a less effective model) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website). Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision floating point the XL model can fit onto 8GB VRAM).
to:
Its original version was powered by the XL-sized (1.5B internal parameters) '''General '''Generative Pre-trained Transformer''' revision 2 (GPT-2), a neural network text predictor expensively ($ 150,000 USD) trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walton (creator of [=AID2=] and a previous incarnation that used a less effective model) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website). Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision floating point the XL model can fit onto 8GB VRAM).
Is there an issue? Send a MessageReason:
None
Changed line(s) 3,6 (click to see context) from:
It is powered by the XL-sized (1.5B internal parameters) '''General Pre-trained Transformer''' revision 2 (now with state caching!), a neural network text predictor expensively ($ 150,000 USD) trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walton (creator of [=AID2=] and a previous incarnation that used a less effective model) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website). Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision floating point the XL model can fit onto 8GB VRAM).
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (any finetuning is done offline at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (any finetuning is done offline at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
to:
Despite appearances
Like most neural network models, GPT-2 is basically a black box. Machine Learning's primary focus is to build and train problem solvers rather than understand the solutions. Better ML architectures require relatively less training effort to perform well. The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is hard-limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=]
The current version of ''AI Dungeon 2'' runs on a new version of the
Changed line(s) 9 (click to see context) from:
Like most neural network models, GPT-2 is basically a black box. Machine Learning's primary focus is to build and train problem solvers rather than understand the solutions. Better ML architectures require relatively less training effort to perform well. The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is hard-limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much (three runs on a swarm of several hundred [=TPUs=]) it is still the best publicly available trained model.
to:
Is there an issue? Send a MessageReason:
reshuffling words, more detail
Changed line(s) 1,10 (click to see context) from:
In a nutshell, AI Dungeon 2 is a generative open-ended text adventure game with some templated scenarios and a limited textual memory to enforce some consistency.
It is powered by the XL-sized (1.5B internal parameters) General Pre-trained Transformer revision 2 (now with state caching!), a neural network text predictor expensively trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walton (creator of [=AID2=]) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website).
Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision the XL model can fit onto 8GB VRAM).
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (it may be trained offline using saved inputs at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it converts word tokens to wavelet space to process them simultaneously, rather than sequentially word-for-word; Attention determines the relative importance of other tokens in the sequence in deciding which word should be generated next).
GPT-2 is basically a black box, Machine Learning's primary focus is to build and train problem solvers rather than understand the solutions. Better ML architectures require relatively less training effort to perform well.
The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is hard-limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much ($ 150,000 USD for three runs on several hundred [=TPUs=]) it is still the best publicly available trained model. And there you have it.
It is powered by the XL-sized (1.5B internal parameters) General Pre-trained Transformer revision 2 (now with state caching!), a neural network text predictor expensively trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walton (creator of [=AID2=]) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website).
Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision the XL model can fit onto 8GB VRAM).
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (it may be trained offline using saved inputs at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it converts word tokens to wavelet space to process them simultaneously, rather than sequentially word-for-word; Attention determines the relative importance of other tokens in the sequence in deciding which word should be generated next).
GPT-2 is basically a black box, Machine Learning's primary focus is to build and train problem solvers rather than understand the solutions. Better ML architectures require relatively less training effort to perform well.
The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is hard-limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much ($ 150,000 USD for three runs on several hundred [=TPUs=]) it is still the best publicly available trained model. And there you have it.
to:
In a nutshell, AI Dungeon 2 is a generative open-ended text adventure game with some templated scenarios and a limited textual memory of prior inputs to enforce some consistency.
It is powered by the XL-sized (1.5B internal parameters)General '''General Pre-trained Transformer Transformer''' revision 2 (now with state caching!), a neural network text predictor expensively ($ 150,000 USD) trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walton (creator of [=AID2=]) [=AID2=] and a previous incarnation that used a less effective model) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website).
website). Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision floating point the XL model can fit onto 8GB VRAM).
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs(it may be trained (any finetuning is done offline using saved inputs at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
time.
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it converts word tokens to wavelet space to process them simultaneously, rather than sequentially word-for-word; Attention determines the relative importance of other tokens in the sequence in deciding which word token should be generated next).
Like most neural network models, GPT-2 is basically a blackbox, box. Machine Learning's primary focus is to build and train problem solvers rather than understand the solutions. Better ML architectures require relatively less training effort to perform well.
well. The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is hard-limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much ($ 150,000 USD for three (three runs on a swarm of several hundred [=TPUs=]) it is still the best publicly available trained model. And there you have it.model.
It is powered by the XL-sized (1.5B internal parameters)
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it converts word tokens to wavelet space to process them simultaneously, rather than sequentially word-for-word; Attention determines the relative importance of other tokens in the sequence in deciding which word token should be generated next).
Like most neural network models, GPT-2 is basically a black
Is there an issue? Send a MessageReason:
reworded the explanation
Changed line(s) 7,8 (click to see context) from:
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it processes word tokens simultaneously after converting them to wavelet space rather than sequentially word-for-word).
to:
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it processes converts word tokens simultaneously after converting them to wavelet space to process them simultaneously, rather than sequentially word-for-word).
word-for-word; Attention determines the relative importance of other tokens in the sequence in deciding which word should be generated next).
Changed line(s) 10 (click to see context) from:
The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much ($ 150,000 USD for three runs on several hundred [=TPUs=]) it is still the best publicly available trained model. And there you have it.
to:
The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is limited hard-limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much ($ 150,000 USD for three runs on several hundred [=TPUs=]) it is still the best publicly available trained model. And there you have it.
Is there an issue? Send a MessageReason:
mistyped Nick Walton (developer)
Changed line(s) 3 (click to see context) from:
It is powered by the XL-sized (1.5B internal parameters) General Pre-trained Transformer revision 2 (now with state caching!), a neural network text predictor expensively trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walden (creator of [=AID2=]) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website).
to:
It is powered by the XL-sized (1.5B internal parameters) General Pre-trained Transformer revision 2 (now with state caching!), a neural network text predictor expensively trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walden Walton (creator of [=AID2=]) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website).
Is there an issue? Send a MessageReason:
Some trivia about the game
Added DiffLines:
In a nutshell, AI Dungeon 2 is a generative open-ended text adventure game with some templated scenarios and a limited textual memory to enforce some consistency.
It is powered by the XL-sized (1.5B internal parameters) General Pre-trained Transformer revision 2 (now with state caching!), a neural network text predictor expensively trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walden (creator of [=AID2=]) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website).
Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision the XL model can fit onto 8GB VRAM).
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (it may be trained offline using saved inputs at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it processes word tokens simultaneously after converting them to wavelet space rather than sequentially word-for-word).
GPT-2 is basically a black box, Machine Learning's primary focus is to build and train problem solvers rather than understand the solutions. Better ML architectures require relatively less training effort to perform well.
The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much ($ 150,000 USD for three runs on several hundred [=TPUs=]) it is still the best publicly available trained model. And there you have it.
It is powered by the XL-sized (1.5B internal parameters) General Pre-trained Transformer revision 2 (now with state caching!), a neural network text predictor expensively trained by [=OpenAI=] on a ~40GB corpus of uncensored Internet text linked with at least 3 karma on Reddit. Nick Walden (creator of [=AID2=]) used transfer learning to finetune this model on the Choose Your Own Adventure format (with text scraped from the eponymous website).
Several forks of the game exist, allowing local play (on CPU or beefy GPU -- with half precision the XL model can fit onto 8GB VRAM).
Despite appearances [=AID2=] is not self-aware and does not learn from player inputs (it may be trained offline using saved inputs at the discretion of the developers). It may follow the "AI speaks directly to you" cliché, causing awe and confusion among players. You may notice the details given vary every time.
The Transformer itself is an Attention-based architecture originally designed by Google Brain researchers (it is named Transformer because it processes word tokens simultaneously after converting them to wavelet space rather than sequentially word-for-word).
GPT-2 is basically a black box, Machine Learning's primary focus is to build and train problem solvers rather than understand the solutions. Better ML architectures require relatively less training effort to perform well.
The GPT-2 architecture itself is no longer state-of-the-art. Transformer-XL and [=XLNet=] are superior in design (T-XL allows for unlimited word tokens, while GPT-2 is limited to 1024; [=XLNet=] also performs bidirectional prediction which is maybe overkill for [=AID2=] given the permutation overhead) but GPT-2 has been trained so much ($ 150,000 USD for three runs on several hundred [=TPUs=]) it is still the best publicly available trained model. And there you have it.