-
Notifications
You must be signed in to change notification settings - Fork 979
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
docs: Update quickstart page to structure things a little more for the novices #1873
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
the documentation changes look good. but this PR is changing many more files - is that intentional?
No, I'll fix this, something went wrong when I made a new branch. Will be cleaned up later this evening. Sorry about that. Update: All cleaned up now! |
2504e59
to
e9d0517
Compare
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
588ae97
to
11b53ac
Compare
@raghotham I incorporated the latest changes and fixed the conflicts. Still gives the same structure (i.e., making a install and setup at the beginning). |
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
1252213
to
5a75727
Compare
Signed-off-by: Francisco Javier Arceo <[email protected]>
@raghotham updated 👍 I'll open the other PR after this is closed to show the smaller demo for getting started. |
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
…ven shorter quickstart to highlight value in as few lines of code as possible Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
Signed-off-by: Francisco Javier Arceo <[email protected]>
@@ -16,3 +16,7 @@ | |||
.hide-title h1 { | |||
display: none; | |||
} | |||
|
|||
h3 { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should it be h2 instead of h3?
Llama Stack config for Ollama. | ||
::::{tab-set} | ||
|
||
:::{tab-item} Using Python |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
should add conda as well
|
||
First lets setup some environment variables and create a local directory to mount into the container’s file system. | ||
```bash | ||
export INFERENCE_MODEL="meta-llama/Llama-3.2-3B-Instruct" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
inference model can be the same as the one used above - we dont need to introduce new names to the same models.
You can reuse the server setup or use the [Llama Stack Client](https://github.com/meta-llama/llama-stack-client-python/). | ||
Note that the client package is already included in the `llama-stack` package. | ||
|
||
### ii. Using the Llama Stack Client |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
there is no ### i.
``` | ||
|
||
## Step 4: Run Basic Inference |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
the rest of the "steps" are just different examples.
source .venv/bin/activate | ||
export INFERENCE_MODEL="llama3.2:3b" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why have the env var here?
docs/source/getting_started/index.md
Outdated
from llama_stack_client import LlamaStackClient | ||
from llama_stack_client import Agent, AgentEventLogger | ||
from termcolor import cprint | ||
from llama_stack.distribution.library_client import LlamaStackAsLibraryClient |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we dont want library as client - let's use server please.
] | ||
vector_db = "faiss" | ||
vector_db_id = "test-vector-db" | ||
model_id = "llama3.2:3b-instruct-fp16" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
wrong model id - it should be llama3.2:3b
client.vector_dbs.register( | ||
provider_id=vector_db, | ||
vector_db_id=vector_db_id, | ||
embedding_model="all-MiniLM-L6-v2", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
embedding model showed up here - can we retrieve it from the distro server?
will make minor fixes after merging this PR |
What does this PR do?
Another doc enhancement for #1818
Summary of changes:
docs/source/distributions/configuration.md
docs/_static/css/my_theme.css
<h3>
elements to set a normal font weight.docs/source/distributions/starting_llama_stack_server.md
##
).docs/source/getting_started/index.md
docs/source/providers/index.md
Let me know if you need further details!
Test Plan
Renders locally, included screenshot.
Documentation
For #1818