Skip to content

Commit

Permalink
Use smaller model for tests with petals
Browse files Browse the repository at this point in the history
  • Loading branch information
mudler committed Dec 20, 2023
1 parent 739ff29 commit 5f56152
Showing 1 changed file with 3 additions and 2 deletions.
5 changes: 3 additions & 2 deletions backend/python/petals/test_petals.py
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,8 @@ def test_load_model(self):
self.setUp()
with grpc.insecure_channel("localhost:50051") as channel:
stub = backend_pb2_grpc.BackendStub(channel)
response = stub.LoadModel(backend_pb2.ModelOptions(Model="petals-team/StableBeluga"))
response = stub.LoadModel(backend_pb2.ModelOptions(Model="bigscience/bloom-560m"))
print(response)
self.assertTrue(response.success)
self.assertEqual(response.message, "Model loaded successfully")
except Exception as err:
Expand All @@ -64,7 +65,7 @@ def test_text(self):
self.setUp()
with grpc.insecure_channel("localhost:50051") as channel:
stub = backend_pb2_grpc.BackendStub(channel)
response = stub.LoadModel(backend_pb2.ModelOptions(Model="petals-team/StableBeluga"))
response = stub.LoadModel(backend_pb2.ModelOptions(Model="bigscience/bloom-560m"))
self.assertTrue(response.success)
req = backend_pb2.PredictOptions(prompt="The capital of France is")
resp = stub.Predict(req)
Expand Down

0 comments on commit 5f56152

Please sign in to comment.