Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Commit bfefa26

Browse filesBrowse files
author
juanroesel
committed
Added simple test to check health of metrics endpoint
1 parent 7ea1675 commit bfefa26
Copy full SHA for bfefa26

File tree

Expand file treeCollapse file tree

1 file changed

+25
-8
lines changed
Filter options
Expand file treeCollapse file tree

1 file changed

+25
-8
lines changed

‎tests/test_llama.py

Copy file name to clipboardExpand all lines: tests/test_llama.py
+25-8Lines changed: 25 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -153,7 +153,8 @@ def mock_kv_cache_seq_add(
153153

154154
def test_llama_patch(mock_llama):
155155
n_ctx = 128
156-
ai_service = "label-suggestions"
156+
ai_service_completion = "test-label-suggestions"
157+
ai_service_streaming = "test-acceptance-criteria"
157158
llama = llama_cpp.Llama(model_path=MODEL, vocab_only=True, n_ctx=n_ctx)
158159

159160
n_vocab = llama_cpp.llama_n_vocab(llama._model.model)
@@ -165,32 +166,32 @@ def test_llama_patch(mock_llama):
165166

166167
## Test basic completion from bos until eos
167168
mock_llama(llama, all_text)
168-
completion = llama.create_completion("", max_tokens=36, ai_service=ai_service)
169+
completion = llama.create_completion("", max_tokens=36, ai_service=ai_service_completion)
169170
assert completion["choices"][0]["text"] == all_text
170171
assert completion["choices"][0]["finish_reason"] == "stop"
171172

172173
## Test basic completion until eos
173174
mock_llama(llama, all_text)
174-
completion = llama.create_completion(text, max_tokens=20, ai_service=ai_service)
175+
completion = llama.create_completion(text, max_tokens=20, ai_service=ai_service_completion)
175176
assert completion["choices"][0]["text"] == output_text
176177
assert completion["choices"][0]["finish_reason"] == "stop"
177178

178179
## Test streaming completion until eos
179180
mock_llama(llama, all_text)
180-
chunks = list(llama.create_completion(text, max_tokens=20, stream=True, ai_service=ai_service))
181+
chunks = list(llama.create_completion(text, max_tokens=20, stream=True, ai_service=ai_service_streaming))
181182
assert "".join(chunk["choices"][0]["text"] for chunk in chunks) == output_text
182183
assert chunks[-1]["choices"][0]["finish_reason"] == "stop"
183184

184185
## Test basic completion until stop sequence
185186
mock_llama(llama, all_text)
186-
completion = llama.create_completion(text, max_tokens=20, stop=["lazy"], ai_service=ai_service)
187+
completion = llama.create_completion(text, max_tokens=20, stop=["lazy"], ai_service=ai_service_completion)
187188
assert completion["choices"][0]["text"] == " jumps over the "
188189
assert completion["choices"][0]["finish_reason"] == "stop"
189190

190191
## Test streaming completion until stop sequence
191192
mock_llama(llama, all_text)
192193
chunks = list(
193-
llama.create_completion(text, max_tokens=20, stream=True, stop=["lazy"], ai_service=ai_service)
194+
llama.create_completion(text, max_tokens=20, stream=True, stop=["lazy"], ai_service=ai_service_streaming)
194195
)
195196
assert (
196197
"".join(chunk["choices"][0]["text"] for chunk in chunks) == " jumps over the "
@@ -199,13 +200,13 @@ def test_llama_patch(mock_llama):
199200

200201
## Test basic completion until length
201202
mock_llama(llama, all_text)
202-
completion = llama.create_completion(text, max_tokens=2, ai_service=ai_service)
203+
completion = llama.create_completion(text, max_tokens=2, ai_service=ai_service_completion)
203204
assert completion["choices"][0]["text"] == " jumps"
204205
assert completion["choices"][0]["finish_reason"] == "length"
205206

206207
## Test streaming completion until length
207208
mock_llama(llama, all_text)
208-
chunks = list(llama.create_completion(text, max_tokens=2, stream=True, ai_service=ai_service))
209+
chunks = list(llama.create_completion(text, max_tokens=2, stream=True, ai_service=ai_service_streaming))
209210
assert "".join(chunk["choices"][0]["text"] for chunk in chunks) == " jumps"
210211
assert chunks[-1]["choices"][0]["finish_reason"] == "length"
211212

@@ -269,6 +270,22 @@ def test_llama_server():
269270
}
270271

271272

273+
def test_metrics_endpoint():
274+
from fastapi.testclient import TestClient
275+
from llama_cpp.server.app import create_app, Settings
276+
277+
settings = Settings(
278+
model=MODEL,
279+
vocab_only=True,
280+
)
281+
app = create_app(settings)
282+
client = TestClient(app)
283+
response = client.get("/metrics")
284+
assert response.status_code == 200
285+
assert "test-label-suggestions" in response.text
286+
assert "test-acceptance-criteria" in response.text
287+
288+
272289
@pytest.mark.parametrize(
273290
"size_and_axis",
274291
[

0 commit comments

Comments
0 (0)
Morty Proxy This is a proxified and sanitized view of the page, visit original site.