Prompt flow Python SDK - build high-quality LLM apps
Project description
Prompt flow
Welcome to join us to make prompt flow better by participating discussions, opening issues, submitting PRs.
Prompt flow is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.
With prompt flow, you will be able to:
- Create and iteratively develop flow
- Create executable flows that link LLMs, prompts, Python code and other tools together.
- Debug and iterate your flows, especially the interaction with LLMs with ease.
- Evaluate flow quality and performance
- Evaluate your flow's quality and performance with larger datasets.
- Integrate the testing and evaluation into your CI/CD system to ensure quality of your flow.
- Streamlined development cycle for production
- Deploy your flow to the serving platform you choose or integrate into your app's code base easily.
- (Optional but highly recommended) Collaborate with your team by leveraging the cloud version of prompt flow in Azure AI.
Installation
Ensure you have a python environment, python=3.9
is recommended.
pip install promptflow promptflow-tools
Quick Start ⚡
Create a chatbot with prompt flow
Run the command to initiate a prompt flow from a chat template, it creates folder named my_chatbot
and generates required files within it:
pf flow init --flow ./my_chatbot --type chat
Setup a connection for your API key
For OpenAI key, establish a connection by running the command, using the openai.yaml
file in the my_chatbot
folder, which stores your OpenAI key:
# Override keys with --set to avoid yaml file changes
pf connection create --file ./my_chatbot/openai.yaml --set api_key=<your_api_key> --name open_ai_connection
For Azure OpenAI key, establish the connection by running the command, using the azure_openai.yaml
file:
pf connection create --file ./my_chatbot/azure_openai.yaml --set api_key=<your_api_key> api_base=<your_api_base> --name open_ai_connection
Chat with your flow
In the my_chatbot
folder, there's a flow.dag.yaml
file that outlines the flow, including inputs/outputs, nodes, connection, and the LLM model, etc
Note that in the
chat
node, we're using a connection namedopen_ai_connection
(specified inconnection
field) and thegpt-35-turbo
model (specified indeployment_name
field). The deployment_name filed is to specify the OpenAI model, or the Azure OpenAI deployment resource.
Interact with your chatbot by running: (press Ctrl + C
to end the session)
pf flow test --flow ./my_chatbot --interactive
Continue to delve deeper into prompt flow.
Release History
1.1.1 (2023.12.1)
Bugs Fixed
- [SDK/CLI] Fix compatibility issue with
semantic-kernel==0.4.0.dev0
andazure-ai-ml==1.12.0
. - [SDK/CLI] Add back workspace information in CLI telemetry.
- [SDK/CLI] Disable the feature to customize user agent in CLI to avoid changes on operation context.
- Fix openai metrics calculator to adapt openai v1.
1.1.0 (2023.11.30)
Features Added
- Add
pfazure flow show/list
to show or list flows from Azure AI. - Display node status in run visualize page graph view.
- Add support for image input and output in prompt flow.
- [SDK/CLI] SDK/CLI will collect telemetry by default, user can use
pf config set telemetry.enabled=false
to opt out. - Add
raise_on_error
for stream run API, by default we raise for failed run. - Flow as function: consume a flow like a function with parameters mapped to flow inputs.
- Enable specifying the default output path for run.
- Use
pf config set run.output_path=<output-path>
to specify, and the run output path will be<output-path>/<run-name>
. - Introduce macro
${flow_directory}
forrun.output_path
in config, which will be replaced with corresponding flow directory. - The flow directory cannot be set as run output path, which means
pf config set run.output_path='${flow_directory}'
is invalid; but you can use child folder, e.g.pf config set run.output_path='${flow_directory}/.runs'
.
- Use
- Support pfazure run create with remote flow.
- For remote workspace flow:
pfazure run create --flow azureml:<flow-name>
- For remote registry flow:
pfazure run create --flow azureml://registries/<registry-name>/models/<flow-name>/versions/<flow-version>
- For remote workspace flow:
- Support set logging level via environment variable
PF_LOGGING_LEVEL
, valid values includesCRITICAL
,ERROR
,WARNING
,INFO
,DEBUG
, default toINFO
. - Remove openai version restrictions
Bugs Fixed
- [SDK/CLI] Fix node test with dict node input will raise "Required input(s) missing".
- [SDK/CLI] Will use run name as display name when display name not specified (used flow folder name before).
- [SDK/CLI] Fix pf flow build created unexpected layer of dist folder
- [SDK/CLI] Fix deploy prompt flow: connections value may be none
Improvements
- Force 'az login' if using azureml connection provider in cli command.
- Add env variable 'PF_NO_INTERACTIVE_LOGIN' to disable interactive login if using azureml connection provider in promptflow sdk.
- Improved CLI invoke time.
- Bump
pydash
upper bound to 8.0.0. - Bump
SQLAlchemy
upper bound to 3.0.0. - Bump
flask
upper bound to 4.0.0,flask-restx
upper bound to 2.0.0. - Bump
ruamel.yaml
upper bound to 1.0.0.
1.0.0 (2023.11.09)
Features Added
- [Executor] Add
enable_kwargs
tag in tools.json for customer python tool. - [SDK/CLI] Support
pfazure flow create
. Create a flow on Azure AI from local flow folder. - [SDK/CLI] Changed column mapping
${run.inputs.xx}
's behavior, it will refer to run's data columns instead of run's inputs columns.
Bugs Fixed
- [SDK/CLI] Keep original format in run output.jsonl.
- [Executor] Fix the bug that raise an error when an aggregation node references a bypassed node
Improvements
- [Executor] Set the outputs of the bypassed nodes as None
0.1.0b8 (2023.10.26)
Features Added
- [Executor] Add average execution time and estimated execution time to batch run logs
- [SDK/CLI] Support
pfazure run archive/restore/update
. - [SDK/CLI] Support custom strong type connection.
- [SDK/CLI] Enable telemetry and won't collect by default, use
pf config set cli.telemetry_enabled=true
to opt in. - [SDK/CLI] Exposed function
from promptflow import load_run
to load run object from local YAML file. - [Executor] Support
ToolProvider
for script tools.
Bugs Fixed
- pf config set:
- Fix bug for workspace
connection.provider=azureml
doesn't work as expected.
- Fix bug for workspace
- [SDK/CLI] Fix the bug that using sdk/cli to submit batch run did not display the log correctly.
- [SDK/CLI] Fix encoding issues when input is non-English with
pf flow test
. - [Executor] Fix the bug can't read file containing "Private Use" unicode character.
- [SDK/CLI] Fix string type data will be converted to integer/float.
- [SDK/CLI] Remove the max rows limitation of loading data.
- [SDK/CLI] Fix the bug --set not taking effect when creating run from file.
Improvements
- [SDK/CLI] Experience improvements in
pf run visualize
page:- Add column status.
- Support opening flow file by clicking run id.
0.1.0b7.post1 (2023.09.28)
Bug Fixed
- Fix extra dependency bug when importing
promptflow
withoutazure-ai-ml
installed.
0.1.0b7 (2023.09.27)
Features Added
- pf flow validate: support validate flow
- pf config set: support set user-level promptflow config.
- Support workspace connection provider, usage:
pf config set connection.provider=azureml://subscriptions/<subscription_id>/resourceGroups/<resource_group>/providers/Microsoft.MachineLearningServices/workspaces/<workspace_name>
- Support workspace connection provider, usage:
- Support override openai connection's model when submitting a flow. For example:
pf run create --flow ./ --data ./data.jsonl --connection llm.model=xxx --column-mapping url='${data.url}'
Bugs Fixed
- [Flow build] Fix flow build file name and environment variable name when connection name contains space.
- Reserve
.promptflow
folder when dump run snapshot. - Read/write log file with encoding specified.
- Avoid inconsistent error message when executor exits abnormally.
- Align inputs & outputs row number in case partial completed run will break
pfazure run show-details
. - Fix bug that failed to parse portal url for run data when the form is an asset id.
- Fix the issue of process hanging for a long time when running the batch run.
Improvements
- [Executor][Internal] Improve error message with more details and actionable information.
- [SDK/CLI]
pf/pfazure run show-details
:- Add
--max-results
option to control the number of results to display. - Add
--all-results
option to display all results.
- Add
- Add validation for azure
PFClient
constructor in case wrong parameter is passed.
0.1.0b6 (2023.09.15)
Features Added
- [promptflow][Feature] Store token metrics in run properties
Bugs Fixed
- Refine error message body for flow_validator.py
- Refine error message body for run_tracker.py
- [Executor][Internal] Add some unit test to improve code coverage of log/metric
- [SDK/CLI] Update portal link to remove flight.
- [Executor][Internal] Improve inputs mapping's error message.
- [API] Resolve warnings/errors of sphinx build
0.1.0b5 (2023.09.08)
Features Added
- pf run visualize: support lineage graph & display name in visualize page
Bugs Fixed
- Add missing requirement
psutil
insetup.py
0.1.0b4 (2023.09.04)
Features added
- Support
pf flow build
commands
0.1.0b3 (2023.08.30)
- Minor bug fixes.
0.1.0b2 (2023.08.29)
- First preview version with major CLI & SDK features.
Features added
- pf flow: init/test/serve/export
- pf run: create/update/stream/list/show/show-details/show-metrics/visualize/archive/restore/export
- pf connection: create/update/show/list/delete
- Azure AI support:
- pfazure run: create/list/stream/show/show-details/show-metrics/visualize
0.1.0b1 (2023.07.20)
- Stub version in Pypi.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distributions
Built Distribution
File details
Details for the file promptflow-1.1.1-py3-none-any.whl
.
File metadata
- Download URL: promptflow-1.1.1-py3-none-any.whl
- Upload date:
- Size: 1.4 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.8.18
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 471b1994ed3e138db4450f8f806867477678c9afd6b097e99c15b2e816ab8abc |
|
MD5 | 464764c11ced52cfcbb843ba60a63464 |
|
BLAKE2b-256 | 082e420bd42864b9f33e78897d96bc4601c1a535db40a79f5cfba702cf505bbd |