Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

NameError: name 'tokenize_and_split_data' is not defined in Python code

I want to divide the data into train_dataset and test_dataset variables. The function tokenize_and_split_data did not work and utilities library did not define. I am working on Python google colab.

import datasets
import tempfile
import logging
import random
import config
import os
import yaml
import time
import torch
import transformers
import pandas as pd
import jsonlines

#from utilities import *
from transformers import AutoTokenizer
from transformers import AutoModelForCausalLM
from transformers import TrainingArguments
from transformers import AutoModelForCausalLM

logger = logging.getLogger(__name__)
global_config = None

model_name = "EleutherAI/pythia-70m"

training_config = {
    "model": {
        "pretrained_name": model_name,
        "max_length" : 2048
    },
    "datasets": {
        "use_hf": use_hf,
        "path": dataset_path
    },
    "verbose": True
}

tokenizer = AutoTokenizer.from_pretrained(model_name)
tokenizer.pad_token = tokenizer.eos_token
train_dataset, test_dataset = tokenize_and_split_data(training_config, tokenizer)

print(train_dataset)
print(test_dataset)

Above, is the code, I cannot install utilities library, and this function tokenize_and_split_data did not defined. Can you help me please.

like image 461
Gha Avatar asked Oct 20 '25 19:10

Gha


1 Answers

If you are running the collab from Lamini on finetuning, there is a python file utilities.py that contains this method. Just recreate this file or copy paste all the methods in a cell.

$ ls
05_Training_lab_student.ipynb  lamini_docs.jsonl    utilities.py
__pycache__                    lamini_docs_3_steps

$ cat utilities.py
like image 115
Kim Noël Avatar answered Oct 22 '25 09:10

Kim Noël



Donate For Us

If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!