| --- |
| license: apache-2.0 |
| base_model: BEE-spoke-data/smol_llama-220M-GQA |
| datasets: |
| - BEE-spoke-data/pypi_clean-deduped |
| - bigcode/the-stack-smol-xl |
| - EleutherAI/proof-pile-2 |
| language: |
| - en |
| tags: |
| - python |
| - codegen |
| - markdown |
| - smol_llama |
| metrics: |
| - accuracy |
| inference: |
| parameters: |
| max_new_tokens: 64 |
| min_new_tokens: 8 |
| do_sample: true |
| epsilon_cutoff: 0.0008 |
| temperature: 0.3 |
| top_p: 0.9 |
| repetition_penalty: 1.02 |
| no_repeat_ngram_size: 8 |
| renormalize_logits: true |
| widget: |
| - text: | |
| def add_numbers(a, b): |
| return |
| example_title: Add Numbers Function |
| - text: | |
| class Car: |
| def __init__(self, make, model): |
| self.make = make |
| self.model = model |
| |
| def display_car(self): |
| example_title: Car Class |
| - text: | |
| import pandas as pd |
| data = {'Name': ['Tom', 'Nick', 'John'], 'Age': [20, 21, 19]} |
| df = pd.DataFrame(data).convert_dtypes() |
| # eda |
| example_title: Pandas DataFrame |
| - text: | |
| def factorial(n): |
| if n == 0: |
| return 1 |
| else: |
| example_title: Factorial Function |
| - text: | |
| def fibonacci(n): |
| if n <= 0: |
| raise ValueError("Incorrect input") |
| elif n == 1: |
| return 0 |
| elif n == 2: |
| return 1 |
| else: |
| example_title: Fibonacci Function |
| - text: | |
| import matplotlib.pyplot as plt |
| import numpy as np |
| x = np.linspace(0, 10, 100) |
| # simple plot |
| example_title: Matplotlib Plot |
| - text: | |
| def reverse_string(s:str) -> str: |
| return |
| example_title: Reverse String Function |
| - text: | |
| def is_palindrome(word:str) -> bool: |
| return |
| example_title: Palindrome Function |
| - text: | |
| def bubble_sort(lst: list): |
| n = len(lst) |
| for i in range(n): |
| for j in range(0, n-i-1): |
| example_title: Bubble Sort Function |
| - text: | |
| def binary_search(arr, low, high, x): |
| if high >= low: |
| mid = (high + low) // 2 |
| if arr[mid] == x: |
| return mid |
| elif arr[mid] > x: |
| example_title: Binary Search Function |
| pipeline_tag: text-generation |
| --- |
| |
| # BEE-spoke-data/beecoder-220M-python |
|
|
|
|
|
|
|
|
| This is `BEE-spoke-data/smol_llama-220M-GQA` fine-tuned for code generation on: |
|
|
| - filtered version of stack-smol-XL |
| - deduped version of 'algebraic stack' from proof-pile-2 |
| - cleaned and deduped pypi (last dataset) |
|
|
| This model (and the base model) were both trained using ctx length 2048. |
|
|
| ## examples |
|
|
| > Example script for inference testing: [here](https://gist.github.com/pszemraj/c7738f664a64b935a558974d23a7aa8c) |
|
|
| It has its limitations at 220M, but seems decent for single-line or docstring generation, and/or being used for speculative decoding for such purposes. |
|
|
|
|
|
|
|  |
|
|
| The screenshot is on CPU on a laptop. |
|
|
| --- |