diff --git a/src/backend/Dockerfile b/src/backend/Dockerfile index f32177b..68e4234 100644 --- a/src/backend/Dockerfile +++ b/src/backend/Dockerfile @@ -6,11 +6,12 @@ COPY lambda_requirements.txt ${LAMBDA_TASK_ROOT} # Install the specified packages RUN pip install --compile --no-cache-dir -r lambda_requirements.txt -# Copy model -COPY trained/mini-copilot ${LAMBDA_TASK_ROOT}/trained/mini-copilot +# Copy old model +COPY trained/gpt2-728 ${LAMBDA_TASK_ROOT}/trained/gpt2-728 -# Copy tokenizer -COPY trained/mini-copilot-tokenizer ${LAMBDA_TASK_ROOT}/trained/mini-copilot-tokenizer +# Copy new model and tokenizer +# COPY trained/mini-copilot ${LAMBDA_TASK_ROOT}/trained/mini-copilot +# COPY trained/mini-copilot-tokenizer ${LAMBDA_TASK_ROOT}/trained/mini-copilot-tokenizer # Copy function code COPY lambda.py ${LAMBDA_TASK_ROOT} diff --git a/src/backend/lambda.py b/src/backend/lambda.py index 6e684a4..4f7b89d 100644 --- a/src/backend/lambda.py +++ b/src/backend/lambda.py @@ -4,9 +4,8 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') -tokenizer = AutoTokenizer.from_pretrained( - "trained/mini-copilot-tokenizer/tokenizer_10M") -model = AutoModelForCausalLM.from_pretrained("trained/mini-copilot/gpt2-large") +tokenizer = AutoTokenizer.from_pretrained("gpt2") +model = AutoModelForCausalLM.from_pretrained("trained/gpt2-728") model.to(device) model.eval()