|
1 | | -import pytest |
| 1 | +import json |
| 2 | +import os |
| 3 | +import time |
2 | 4 |
|
3 | | -from .rest_api_utils import ( # CREATE_FINE_TUNE_DI_BATCH_JOB_BUNDLE_REQUEST, CREATE_FINE_TUNE_REQUEST, USER_ID_0, cancel_fine_tune_by_id, create_docker_image_batch_job_bundle, create_fine_tune, get_fine_tune_by_id, |
| 5 | +import boto3 |
| 6 | +import smart_open |
| 7 | + |
| 8 | +from .rest_api_utils import ( |
| 9 | + CREATE_FINE_TUNE_DI_BATCH_JOB_BUNDLE_REQUEST, |
| 10 | + CREATE_FINE_TUNE_REQUEST, |
4 | 11 | USER_ID_0, |
| 12 | + cancel_fine_tune_by_id, |
| 13 | + create_docker_image_batch_job_bundle, |
| 14 | + create_fine_tune, |
| 15 | + get_fine_tune_by_id, |
5 | 16 | list_fine_tunes, |
6 | 17 | ) |
7 | 18 |
|
| 19 | +MAX_RETRIES = 10 |
| 20 | + |
8 | 21 |
|
9 | | -@pytest.mark.skip(reason="test doesn't currently work, needs to be implemented correctly") |
10 | 22 | def test_fine_tunes() -> None: |
11 | | - # TODO: get this test to work (move LLM fine tune repository to database rather than in S3) |
| 23 | + di_batch_job_id = create_docker_image_batch_job_bundle( |
| 24 | + CREATE_FINE_TUNE_DI_BATCH_JOB_BUNDLE_REQUEST, USER_ID_0 |
| 25 | + )["docker_image_batch_job_bundle_id"] |
| 26 | + data = { |
| 27 | + "test_base_model-lora": { |
| 28 | + "docker_image_batch_job_bundle_id": di_batch_job_id, |
| 29 | + "launch_bundle_config": {}, |
| 30 | + "launch_endpoint_config": {}, |
| 31 | + "default_hparams": {}, |
| 32 | + "required_params": [], |
| 33 | + } |
| 34 | + } |
12 | 35 |
|
13 | | - # di_batch_job_id = create_docker_image_batch_job_bundle( |
14 | | - # CREATE_FINE_TUNE_DI_BATCH_JOB_BUNDLE_REQUEST, USER_ID_0 |
15 | | - # )["docker_image_batch_job_bundle_id"] |
| 36 | + if os.getenv("CIRCLECI") == "true": |
| 37 | + session = boto3.Session() |
| 38 | + aws_s3_bucket = os.getenv("CIRCLECI_AWS_S3_BUCKET") |
| 39 | + client = session.client("s3") |
| 40 | + with smart_open.open( |
| 41 | + f"s3://{aws_s3_bucket}/fine_tune_repository", |
| 42 | + "w", |
| 43 | + transport_params={"client": client}, |
| 44 | + ) as f: |
| 45 | + json.dump(data, f) |
16 | 46 |
|
17 | | - # create_response = create_fine_tune(CREATE_FINE_TUNE_REQUEST, USER_ID_0) |
18 | | - # fine_tune_id = create_response["id"] |
| 47 | + create_response = create_fine_tune(CREATE_FINE_TUNE_REQUEST, USER_ID_0) |
| 48 | + fine_tune_id = create_response["id"] |
19 | 49 |
|
20 | | - # get_response = get_fine_tune_by_id(fine_tune_id, USER_ID_0) |
21 | | - # assert get_response["id"] == fine_tune_id |
| 50 | + get_response = get_fine_tune_by_id(fine_tune_id, USER_ID_0) |
| 51 | + num_retries = 0 |
| 52 | + while get_response["status"] not in ["SUCCESS", "FAILURE"]: |
| 53 | + if num_retries >= MAX_RETRIES: |
| 54 | + raise Exception("Fine tune job did not complete in time.") |
| 55 | + num_retries += 1 |
| 56 | + get_response = get_fine_tune_by_id(fine_tune_id, USER_ID_0) |
| 57 | + time.sleep(10) |
| 58 | + assert get_response["id"] == fine_tune_id |
| 59 | + assert get_response["status"] == "SUCCESS" |
22 | 60 |
|
23 | | - # list_response_0_before = list_fine_tunes(USER_ID_0) |
24 | | - # num_jobs = len(list_response_0_before["jobs"]) |
25 | | - # assert num_jobs >= 1 |
| 61 | + list_response_0_before = list_fine_tunes(USER_ID_0) |
| 62 | + num_jobs = len(list_response_0_before["jobs"]) |
| 63 | + assert num_jobs >= 1 |
26 | 64 |
|
27 | | - list_response_1 = list_fine_tunes(USER_ID_0) |
28 | | - assert len(list_response_1["jobs"]) == 0 |
| 65 | + cancel_fine_tune_by_id(fine_tune_id, USER_ID_0) |
29 | 66 |
|
30 | | - # list_response_0_after = list_fine_tunes(USER_ID_0) |
31 | | - # assert len(list_response_0_after["jobs"]) == num_jobs - 1 |
| 67 | + list_response_0_after = list_fine_tunes(USER_ID_0) |
| 68 | + assert len(list_response_0_after["jobs"]) == num_jobs - 1 |
0 commit comments