Update README.md
Browse files
README.md
CHANGED
|
@@ -4,6 +4,21 @@ license: mit
|
|
| 4 |
|
| 5 |
# PyAutoCode: GPT-2 based Python auto-code.
|
| 6 |
|
| 7 |
-
PyAutoCode is a cut-down python autosuggestion built on **GPT-2** (motivation: GPyT) model. This baby model is not **"fine-tuned"** yet therefore, I highly recommend not to use it in a production environment or incorporate PyAutoCode in any of your projects. It has been trained on **112GB** of Python data sourced from the best crowdsource platform ever -- **GitHub**.
|
| 8 |
|
|
|
|
| 9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
# PyAutoCode: GPT-2 based Python auto-code.
|
| 6 |
|
| 7 |
+
PyAutoCode is a cut-down python autosuggestion built on **GPT-2** *(motivation: GPyT)* model. This baby model *(trained only up to 3 epochs)* is not **"fine-tuned"** yet therefore, I highly recommend not to use it in a production environment or incorporate PyAutoCode in any of your projects. It has been trained on **112GB** of Python data sourced from the best crowdsource platform ever -- **GitHub**.
|
| 8 |
|
| 9 |
+
*NOTE: Increased training and fine tuning would be highly appreciated and I firmly believe that it would improve the ability of PyAutoCode significantly.*
|
| 10 |
|
| 11 |
+
## Some Model Features
|
| 12 |
+
|
| 13 |
+
- Built on *GPT-2*
|
| 14 |
+
- Tokenized with *ByteLevelBPETokenizer*
|
| 15 |
+
- Data Sourced from *GitHub (almost 5 consecutive days of latest Python repositories)*
|
| 16 |
+
- Makes use of *GPTLMHeadModel* and *DataCollatorForLanguageModelling* for training
|
| 17 |
+
|
| 18 |
+
## Usage
|
| 19 |
+
|
| 20 |
+
You can use my model too!. Here's a quick tour of how you can achieve this:
|
| 21 |
+
|
| 22 |
+
```python
|
| 23 |
+
|
| 24 |
+
```
|