| | --- |
| | language: code |
| | thumbnail: |
| | --- |
| | |
| | # CodeBERTaPy |
| |
|
| | CodeBERTaPy is a RoBERTa-like model trained on the [CodeSearchNet](https://github.blog/2019-09-26-introducing-the-codesearchnet-challenge/) dataset from GitHub for `python` by [Manuel Romero](https://twitter.com/mrm8488) |
| |
|
| | The **tokenizer** is a Byte-level BPE tokenizer trained on the corpus using Hugging Face `tokenizers`. |
| |
|
| | Because it is trained on a corpus of code (vs. natural language), it encodes the corpus efficiently (the sequences are between 33% to 50% shorter, compared to the same corpus tokenized by gpt2/roberta). |
| |
|
| | The (small) **model** is a 6-layer, 84M parameters, RoBERTa-like Transformer model โ thatโs the same number of layers & heads as DistilBERT โ initialized from the default initialization settings and trained from scratch on the full `python` corpus for 4 epochs. |
| |
|
| | ## Quick start: masked language modeling prediction |
| |
|
| | ```python |
| | PYTHON_CODE = """ |
| | fruits = ['apples', 'bananas', 'oranges'] |
| | for idx, <mask> in enumerate(fruits): |
| | print("index is %d and value is %s" % (idx, val)) |
| | """.lstrip() |
| | ``` |
| |
|
| | ### Does the model know how to complete simple Python code? |
| |
|
| | ```python |
| | from transformers import pipeline |
| | |
| | fill_mask = pipeline( |
| | "fill-mask", |
| | model="mrm8488/CodeBERTaPy", |
| | tokenizer="mrm8488/CodeBERTaPy" |
| | ) |
| | |
| | fill_mask(PYTHON_CODE) |
| | |
| | ## Top 5 predictions: |
| | |
| | 'val' # prob 0.980728805065155 |
| | 'value' |
| | 'idx' |
| | ',val' |
| | '_' |
| | ``` |
| |
|
| | ### Yes! That was easy ๐ Let's try with another Flask like example |
| |
|
| | ```python |
| | PYTHON_CODE2 = """ |
| | @app.route('/<name>') |
| | def hello_name(name): |
| | return "Hello {}!".format(<mask>) |
| | |
| | if __name__ == '__main__': |
| | app.run() |
| | """.lstrip() |
| | |
| | |
| | fill_mask(PYTHON_CODE2) |
| | |
| | ## Top 5 predictions: |
| | |
| | 'name' # prob 0.9961813688278198 |
| | ' name' |
| | 'url' |
| | 'description' |
| | 'self' |
| | ``` |
| |
|
| | ### Yeah! It works ๐ Let's try with another Tensorflow/Keras like example |
| |
|
| | ```python |
| | PYTHON_CODE3=""" |
| | model = keras.Sequential([ |
| | keras.layers.Flatten(input_shape=(28, 28)), |
| | keras.layers.<mask>(128, activation='relu'), |
| | keras.layers.Dense(10, activation='softmax') |
| | ]) |
| | """.lstrip() |
| | |
| | |
| | fill_mask(PYTHON_CODE3) |
| | |
| | ## Top 5 predictions: |
| | |
| | 'Dense' # prob 0.4482928514480591 |
| | 'relu' |
| | 'Flatten' |
| | 'Activation' |
| | 'Conv' |
| | ``` |
| |
|
| | > Great! ๐ |
| |
|
| | ## This work is heavily inspired on [CodeBERTa](https://github.com/huggingface/transformers/blob/master/model_cards/huggingface/CodeBERTa-small-v1/README.md) by huggingface team |
| |
|
| | <br> |
| |
|
| | ## CodeSearchNet citation |
| |
|
| | <details> |
| |
|
| | ```bibtex |
| | @article{husain_codesearchnet_2019, |
| | title = {{CodeSearchNet} {Challenge}: {Evaluating} the {State} of {Semantic} {Code} {Search}}, |
| | shorttitle = {{CodeSearchNet} {Challenge}}, |
| | url = {http://arxiv.org/abs/1909.09436}, |
| | urldate = {2020-03-12}, |
| | journal = {arXiv:1909.09436 [cs, stat]}, |
| | author = {Husain, Hamel and Wu, Ho-Hsiang and Gazit, Tiferet and Allamanis, Miltiadis and Brockschmidt, Marc}, |
| | month = sep, |
| | year = {2019}, |
| | note = {arXiv: 1909.09436}, |
| | } |
| | ``` |
| |
|
| | </details> |
| |
|
| | > Created by [Manuel Romero/@mrm8488](https://twitter.com/mrm8488) |
| |
|
| | > Made with <span style="color: #e25555;">♥</span> in Spain |
| |
|