Skip to content

Commit f1a093b

Browse files
chore(docs): add missing descriptions
1 parent f014d4c commit f1a093b

File tree

22 files changed

+566
-0
lines changed

22 files changed

+566
-0
lines changed

src/llama_stack_client/_client.py

Lines changed: 246 additions & 0 deletions
Large diffs are not rendered by default.

src/llama_stack_client/resources/alpha/alpha.py

Lines changed: 66 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,9 @@ def benchmarks(self) -> BenchmarksResource:
6565

6666
@cached_property
6767
def eval(self) -> EvalResource:
68+
"""
69+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
70+
"""
6871
return EvalResource(self._client)
6972

7073
@cached_property
@@ -73,6 +76,14 @@ def admin(self) -> AdminResource:
7376

7477
@cached_property
7578
def inference(self) -> InferenceResource:
79+
"""
80+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
81+
82+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
83+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
84+
- Embedding models: these models generate embeddings to be used for semantic search.
85+
- Rerank models: these models reorder the documents based on their relevance to a query.
86+
"""
7687
return InferenceResource(self._client)
7788

7889
@cached_property
@@ -106,6 +117,9 @@ def benchmarks(self) -> AsyncBenchmarksResource:
106117

107118
@cached_property
108119
def eval(self) -> AsyncEvalResource:
120+
"""
121+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
122+
"""
109123
return AsyncEvalResource(self._client)
110124

111125
@cached_property
@@ -114,6 +128,14 @@ def admin(self) -> AsyncAdminResource:
114128

115129
@cached_property
116130
def inference(self) -> AsyncInferenceResource:
131+
"""
132+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
133+
134+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
135+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
136+
- Embedding models: these models generate embeddings to be used for semantic search.
137+
- Rerank models: these models reorder the documents based on their relevance to a query.
138+
"""
117139
return AsyncInferenceResource(self._client)
118140

119141
@cached_property
@@ -150,6 +172,9 @@ def benchmarks(self) -> BenchmarksResourceWithRawResponse:
150172

151173
@cached_property
152174
def eval(self) -> EvalResourceWithRawResponse:
175+
"""
176+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
177+
"""
153178
return EvalResourceWithRawResponse(self._alpha.eval)
154179

155180
@cached_property
@@ -158,6 +183,14 @@ def admin(self) -> AdminResourceWithRawResponse:
158183

159184
@cached_property
160185
def inference(self) -> InferenceResourceWithRawResponse:
186+
"""
187+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
188+
189+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
190+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
191+
- Embedding models: these models generate embeddings to be used for semantic search.
192+
- Rerank models: these models reorder the documents based on their relevance to a query.
193+
"""
161194
return InferenceResourceWithRawResponse(self._alpha.inference)
162195

163196

@@ -175,6 +208,9 @@ def benchmarks(self) -> AsyncBenchmarksResourceWithRawResponse:
175208

176209
@cached_property
177210
def eval(self) -> AsyncEvalResourceWithRawResponse:
211+
"""
212+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
213+
"""
178214
return AsyncEvalResourceWithRawResponse(self._alpha.eval)
179215

180216
@cached_property
@@ -183,6 +219,14 @@ def admin(self) -> AsyncAdminResourceWithRawResponse:
183219

184220
@cached_property
185221
def inference(self) -> AsyncInferenceResourceWithRawResponse:
222+
"""
223+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
224+
225+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
226+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
227+
- Embedding models: these models generate embeddings to be used for semantic search.
228+
- Rerank models: these models reorder the documents based on their relevance to a query.
229+
"""
186230
return AsyncInferenceResourceWithRawResponse(self._alpha.inference)
187231

188232

@@ -200,6 +244,9 @@ def benchmarks(self) -> BenchmarksResourceWithStreamingResponse:
200244

201245
@cached_property
202246
def eval(self) -> EvalResourceWithStreamingResponse:
247+
"""
248+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
249+
"""
203250
return EvalResourceWithStreamingResponse(self._alpha.eval)
204251

205252
@cached_property
@@ -208,6 +255,14 @@ def admin(self) -> AdminResourceWithStreamingResponse:
208255

209256
@cached_property
210257
def inference(self) -> InferenceResourceWithStreamingResponse:
258+
"""
259+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
260+
261+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
262+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
263+
- Embedding models: these models generate embeddings to be used for semantic search.
264+
- Rerank models: these models reorder the documents based on their relevance to a query.
265+
"""
211266
return InferenceResourceWithStreamingResponse(self._alpha.inference)
212267

213268

@@ -225,6 +280,9 @@ def benchmarks(self) -> AsyncBenchmarksResourceWithStreamingResponse:
225280

226281
@cached_property
227282
def eval(self) -> AsyncEvalResourceWithStreamingResponse:
283+
"""
284+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
285+
"""
228286
return AsyncEvalResourceWithStreamingResponse(self._alpha.eval)
229287

230288
@cached_property
@@ -233,4 +291,12 @@ def admin(self) -> AsyncAdminResourceWithStreamingResponse:
233291

234292
@cached_property
235293
def inference(self) -> AsyncInferenceResourceWithStreamingResponse:
294+
"""
295+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
296+
297+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
298+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
299+
- Embedding models: these models generate embeddings to be used for semantic search.
300+
- Rerank models: these models reorder the documents based on their relevance to a query.
301+
"""
236302
return AsyncInferenceResourceWithStreamingResponse(self._alpha.inference)

src/llama_stack_client/resources/alpha/eval/eval.py

Lines changed: 26 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -45,8 +45,15 @@
4545

4646

4747
class EvalResource(SyncAPIResource):
48+
"""
49+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
50+
"""
51+
4852
@cached_property
4953
def jobs(self) -> JobsResource:
54+
"""
55+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
56+
"""
5057
return JobsResource(self._client)
5158

5259
@cached_property
@@ -254,8 +261,15 @@ def run_eval_alpha(
254261

255262

256263
class AsyncEvalResource(AsyncAPIResource):
264+
"""
265+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
266+
"""
267+
257268
@cached_property
258269
def jobs(self) -> AsyncJobsResource:
270+
"""
271+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
272+
"""
259273
return AsyncJobsResource(self._client)
260274

261275
@cached_property
@@ -483,6 +497,9 @@ def __init__(self, eval: EvalResource) -> None:
483497

484498
@cached_property
485499
def jobs(self) -> JobsResourceWithRawResponse:
500+
"""
501+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
502+
"""
486503
return JobsResourceWithRawResponse(self._eval.jobs)
487504

488505

@@ -505,6 +522,9 @@ def __init__(self, eval: AsyncEvalResource) -> None:
505522

506523
@cached_property
507524
def jobs(self) -> AsyncJobsResourceWithRawResponse:
525+
"""
526+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
527+
"""
508528
return AsyncJobsResourceWithRawResponse(self._eval.jobs)
509529

510530

@@ -527,6 +547,9 @@ def __init__(self, eval: EvalResource) -> None:
527547

528548
@cached_property
529549
def jobs(self) -> JobsResourceWithStreamingResponse:
550+
"""
551+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
552+
"""
530553
return JobsResourceWithStreamingResponse(self._eval.jobs)
531554

532555

@@ -549,4 +572,7 @@ def __init__(self, eval: AsyncEvalResource) -> None:
549572

550573
@cached_property
551574
def jobs(self) -> AsyncJobsResourceWithStreamingResponse:
575+
"""
576+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
577+
"""
552578
return AsyncJobsResourceWithStreamingResponse(self._eval.jobs)

src/llama_stack_client/resources/alpha/eval/jobs.py

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,10 @@
2727

2828

2929
class JobsResource(SyncAPIResource):
30+
"""
31+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
32+
"""
33+
3034
@cached_property
3135
def with_raw_response(self) -> JobsResourceWithRawResponse:
3236
"""
@@ -157,6 +161,10 @@ def status(
157161

158162

159163
class AsyncJobsResource(AsyncAPIResource):
164+
"""
165+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
166+
"""
167+
160168
@cached_property
161169
def with_raw_response(self) -> AsyncJobsResourceWithRawResponse:
162170
"""

src/llama_stack_client/resources/alpha/inference.py

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -31,6 +31,15 @@
3131

3232

3333
class InferenceResource(SyncAPIResource):
34+
"""
35+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
36+
37+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
38+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
39+
- Embedding models: these models generate embeddings to be used for semantic search.
40+
- Rerank models: these models reorder the documents based on their relevance to a query.
41+
"""
42+
3443
@cached_property
3544
def with_raw_response(self) -> InferenceResourceWithRawResponse:
3645
"""
@@ -109,6 +118,15 @@ def rerank(
109118

110119

111120
class AsyncInferenceResource(AsyncAPIResource):
121+
"""
122+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
123+
124+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
125+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
126+
- Embedding models: these models generate embeddings to be used for semantic search.
127+
- Rerank models: these models reorder the documents based on their relevance to a query.
128+
"""
129+
112130
@cached_property
113131
def with_raw_response(self) -> AsyncInferenceResourceWithRawResponse:
114132
"""

src/llama_stack_client/resources/batches.py

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -35,6 +35,15 @@
3535

3636

3737
class BatchesResource(SyncAPIResource):
38+
"""
39+
The API is designed to allow use of openai client libraries for seamless integration.
40+
41+
This API provides the following extensions:
42+
- idempotent batch creation
43+
44+
Note: This API is currently under active development and may undergo changes.
45+
"""
46+
3847
@cached_property
3948
def with_raw_response(self) -> BatchesResourceWithRawResponse:
4049
"""
@@ -228,6 +237,15 @@ def cancel(
228237

229238

230239
class AsyncBatchesResource(AsyncAPIResource):
240+
"""
241+
The API is designed to allow use of openai client libraries for seamless integration.
242+
243+
This API provides the following extensions:
244+
- idempotent batch creation
245+
246+
Note: This API is currently under active development and may undergo changes.
247+
"""
248+
231249
@cached_property
232250
def with_raw_response(self) -> AsyncBatchesResourceWithRawResponse:
233251
"""

src/llama_stack_client/resources/chat/chat.py

Lines changed: 48 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,14 @@
2525
class ChatResource(SyncAPIResource):
2626
@cached_property
2727
def completions(self) -> CompletionsResource:
28+
"""
29+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
30+
31+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
32+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
33+
- Embedding models: these models generate embeddings to be used for semantic search.
34+
- Rerank models: these models reorder the documents based on their relevance to a query.
35+
"""
2836
return CompletionsResource(self._client)
2937

3038
@cached_property
@@ -50,6 +58,14 @@ def with_streaming_response(self) -> ChatResourceWithStreamingResponse:
5058
class AsyncChatResource(AsyncAPIResource):
5159
@cached_property
5260
def completions(self) -> AsyncCompletionsResource:
61+
"""
62+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
63+
64+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
65+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
66+
- Embedding models: these models generate embeddings to be used for semantic search.
67+
- Rerank models: these models reorder the documents based on their relevance to a query.
68+
"""
5369
return AsyncCompletionsResource(self._client)
5470

5571
@cached_property
@@ -78,6 +94,14 @@ def __init__(self, chat: ChatResource) -> None:
7894

7995
@cached_property
8096
def completions(self) -> CompletionsResourceWithRawResponse:
97+
"""
98+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
99+
100+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
101+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
102+
- Embedding models: these models generate embeddings to be used for semantic search.
103+
- Rerank models: these models reorder the documents based on their relevance to a query.
104+
"""
81105
return CompletionsResourceWithRawResponse(self._chat.completions)
82106

83107

@@ -87,6 +111,14 @@ def __init__(self, chat: AsyncChatResource) -> None:
87111

88112
@cached_property
89113
def completions(self) -> AsyncCompletionsResourceWithRawResponse:
114+
"""
115+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
116+
117+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
118+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
119+
- Embedding models: these models generate embeddings to be used for semantic search.
120+
- Rerank models: these models reorder the documents based on their relevance to a query.
121+
"""
90122
return AsyncCompletionsResourceWithRawResponse(self._chat.completions)
91123

92124

@@ -96,6 +128,14 @@ def __init__(self, chat: ChatResource) -> None:
96128

97129
@cached_property
98130
def completions(self) -> CompletionsResourceWithStreamingResponse:
131+
"""
132+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
133+
134+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
135+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
136+
- Embedding models: these models generate embeddings to be used for semantic search.
137+
- Rerank models: these models reorder the documents based on their relevance to a query.
138+
"""
99139
return CompletionsResourceWithStreamingResponse(self._chat.completions)
100140

101141

@@ -105,4 +145,12 @@ def __init__(self, chat: AsyncChatResource) -> None:
105145

106146
@cached_property
107147
def completions(self) -> AsyncCompletionsResourceWithStreamingResponse:
148+
"""
149+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
150+
151+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
152+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
153+
- Embedding models: these models generate embeddings to be used for semantic search.
154+
- Rerank models: these models reorder the documents based on their relevance to a query.
155+
"""
108156
return AsyncCompletionsResourceWithStreamingResponse(self._chat.completions)

0 commit comments

Comments
 (0)