U
    ~fh                     @  s  d dl mZ d dlmZmZmZ d dlmZ d dlZddl	m
Z
 ddlmZmZmZmZmZ ddlmZmZ dd	lmZ dd
lmZmZ ddlmZmZ ddlmZmZ ddlm Z m!Z!m"Z" ddl#m$Z$m%Z% ddl&m'Z' ddl(m)Z) ddl*m+Z+ ddl,m-Z- ddl.m/Z/ ddl0m1Z1 ddl2m3Z3 ddgZ4G dd deZ5G dd deZ6G dd dZ7G dd dZ8G dd dZ9G d d! d!Z:dS )"    )annotations)UnionIterableOptional)LiteralN   )_legacy_response)	NOT_GIVENBodyQueryHeadersNotGiven)maybe_transformasync_maybe_transform)cached_property)SyncAPIResourceAsyncAPIResource)to_streamed_response_wrapper"async_to_streamed_response_wrapper)SyncCursorPageAsyncCursorPage)assistant_list_paramsassistant_create_paramsassistant_update_params)AsyncPaginatormake_request_options)	Assistant)	ChatModel)AssistantDeleted)Metadata)ReasoningEffort)AssistantToolParam)"AssistantResponseFormatOptionParam
AssistantsAsyncAssistantsc                   @  sF  e Zd ZeddddZeddddZeeeeeeeeeeddded	d
ddddddddddddddddddZdddeddddddddddZeeeeeeeeeeedddedddddd ddddd!dddddddd"d#d$Z	eeeeddded%d&d&d'd(ddddd)d*	d+d,Z
dddeddddddd-dd.d/ZdS )0r#   AssistantsWithRawResponsereturnc                 C  s   t | S a  
        This property can be used as a prefix for any HTTP method call to return
        the raw response object instead of the parsed content.

        For more information, see https://www.github.com/openai/openai-python#accessing-raw-response-data-eg-headers
        )r%   self r+   D/tmp/pip-unpacked-wheel-rjiviquo/openai/resources/beta/assistants.pywith_raw_response#   s    zAssistants.with_raw_responseAssistantsWithStreamingResponsec                 C  s   t | S z
        An alternative to `.with_raw_response` that doesn't eagerly read the response body.

        For more information, see https://www.github.com/openai/openai-python#with_streaming_response
        )r.   r)   r+   r+   r,   with_streaming_response-   s    z"Assistants.with_streaming_responseNdescriptioninstructionsmetadatanamereasoning_effortresponse_formattemperaturetool_resourcestoolstop_pextra_headersextra_query
extra_bodytimeoutUnion[str, ChatModel]Optional[str] | NotGivenOptional[Metadata] | NotGiven$Optional[ReasoningEffort] | NotGiven7Optional[AssistantResponseFormatOptionParam] | NotGivenOptional[float] | NotGiven:Optional[assistant_create_params.ToolResources] | NotGiven'Iterable[AssistantToolParam] | NotGivenHeaders | NoneQuery | NoneBody | None'float | httpx.Timeout | None | NotGivenr   modelr2   r3   r4   r5   r6   r7   r8   r9   r:   r;   r<   r=   r>   r?   r'   c                C  sN   ddi|pi }| j dt|||||||||	|
|dtjt||||dtdS )  
        Create an assistant with a model and instructions.

        Args:
          model: ID of the model to use. You can use the
              [List models](https://platform.openai.com/docs/api-reference/models/list) API to
              see all of your available models, or see our
              [Model overview](https://platform.openai.com/docs/models) for descriptions of
              them.

          description: The description of the assistant. The maximum length is 512 characters.

          instructions: The system instructions that the assistant uses. The maximum length is 256,000
              characters.

          metadata: Set of 16 key-value pairs that can be attached to an object. This can be useful
              for storing additional information about the object in a structured format, and
              querying for objects via API or the dashboard.

              Keys are strings with a maximum length of 64 characters. Values are strings with
              a maximum length of 512 characters.

          name: The name of the assistant. The maximum length is 256 characters.

          reasoning_effort: **o-series models only**

              Constrains effort on reasoning for
              [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
              supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
              result in faster responses and fewer tokens used on reasoning in a response.

          response_format: Specifies the format that the model must output. Compatible with
              [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
              [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
              and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.

              Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
              Outputs which ensures the model will match your supplied JSON schema. Learn more
              in the
              [Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).

              Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
              message the model generates is valid JSON.

              **Important:** when using JSON mode, you **must** also instruct the model to
              produce JSON yourself via a system or user message. Without this, the model may
              generate an unending stream of whitespace until the generation reaches the token
              limit, resulting in a long-running and seemingly "stuck" request. Also note that
              the message content may be partially cut off if `finish_reason="length"`, which
              indicates the generation exceeded `max_tokens` or the conversation exceeded the
              max context length.

          temperature: What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
              make the output more random, while lower values like 0.2 will make it more
              focused and deterministic.

          tool_resources: A set of resources that are used by the assistant's tools. The resources are
              specific to the type of tool. For example, the `code_interpreter` tool requires
              a list of file IDs, while the `file_search` tool requires a list of vector store
              IDs.

          tools: A list of tool enabled on the assistant. There can be a maximum of 128 tools per
              assistant. Tools can be of types `code_interpreter`, `file_search`, or
              `function`.

          top_p: An alternative to sampling with temperature, called nucleus sampling, where the
              model considers the results of the tokens with top_p probability mass. So 0.1
              means only the tokens comprising the top 10% probability mass are considered.

              We generally recommend altering this or temperature but not both.

          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        OpenAI-Betaassistants=v2/assistantsrM   r2   r3   r4   r5   r6   r7   r8   r9   r:   r;   r<   r=   r>   r?   bodyoptionscast_to)_postr   r   AssistantCreateParamsr   r   r*   rM   r2   r3   r4   r5   r6   r7   r8   r9   r:   r;   r<   r=   r>   r?   r+   r+   r,   create6   s4    d   zAssistants.createrS   strassistant_idr<   r=   r>   r?   r'   c                C  sD   |st d|ddi|pi }| jd| t||||dtdS )K  
        Retrieves an assistant.

        Args:
          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        ;Expected a non-empty value for `assistant_id` but received rO   rP   /assistants/rS   rV   rW   
ValueError_getr   r   r*   r^   r<   r=   r>   r?   r+   r+   r,   retrieve   s       zAssistants.retriever2   r3   r4   rM   r5   r6   r7   r8   r9   r:   r;   r<   r=   r>   r?     Union[str, Literal[('gpt-4.1', 'gpt-4.1-mini', 'gpt-4.1-nano', 'gpt-4.1-2025-04-14', 'gpt-4.1-mini-2025-04-14', 'gpt-4.1-nano-2025-04-14', 'o3-mini', 'o3-mini-2025-01-31', 'o1', 'o1-2024-12-17', 'gpt-4o', 'gpt-4o-2024-11-20', 'gpt-4o-2024-08-06', 'gpt-4o-2024-05-13', 'gpt-4o-mini', 'gpt-4o-mini-2024-07-18', 'gpt-4.5-preview', 'gpt-4.5-preview-2025-02-27', 'gpt-4-turbo', 'gpt-4-turbo-2024-04-09', 'gpt-4-0125-preview', 'gpt-4-turbo-preview', 'gpt-4-1106-preview', 'gpt-4-vision-preview', 'gpt-4', 'gpt-4-0314', 'gpt-4-0613', 'gpt-4-32k', 'gpt-4-32k-0314', 'gpt-4-32k-0613', 'gpt-3.5-turbo', 'gpt-3.5-turbo-16k', 'gpt-3.5-turbo-0613', 'gpt-3.5-turbo-1106', 'gpt-3.5-turbo-0125', 'gpt-3.5-turbo-16k-0613')]] | NotGiven:Optional[assistant_update_params.ToolResources] | NotGivenr^   r2   r3   r4   rM   r5   r6   r7   r8   r9   r:   r;   r<   r=   r>   r?   r'   c                C  sf   |st d|ddi|pi }| jd| t||||||||	|
||dtjt||||dtdS )  Modifies an assistant.

        Args:
          description: The description of the assistant.

        The maximum length is 512 characters.

          instructions: The system instructions that the assistant uses. The maximum length is 256,000
              characters.

          metadata: Set of 16 key-value pairs that can be attached to an object. This can be useful
              for storing additional information about the object in a structured format, and
              querying for objects via API or the dashboard.

              Keys are strings with a maximum length of 64 characters. Values are strings with
              a maximum length of 512 characters.

          model: ID of the model to use. You can use the
              [List models](https://platform.openai.com/docs/api-reference/models/list) API to
              see all of your available models, or see our
              [Model overview](https://platform.openai.com/docs/models) for descriptions of
              them.

          name: The name of the assistant. The maximum length is 256 characters.

          reasoning_effort: **o-series models only**

              Constrains effort on reasoning for
              [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
              supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
              result in faster responses and fewer tokens used on reasoning in a response.

          response_format: Specifies the format that the model must output. Compatible with
              [GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
              [GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
              and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.

              Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
              Outputs which ensures the model will match your supplied JSON schema. Learn more
              in the
              [Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).

              Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
              message the model generates is valid JSON.

              **Important:** when using JSON mode, you **must** also instruct the model to
              produce JSON yourself via a system or user message. Without this, the model may
              generate an unending stream of whitespace until the generation reaches the token
              limit, resulting in a long-running and seemingly "stuck" request. Also note that
              the message content may be partially cut off if `finish_reason="length"`, which
              indicates the generation exceeded `max_tokens` or the conversation exceeded the
              max context length.

          temperature: What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
              make the output more random, while lower values like 0.2 will make it more
              focused and deterministic.

          tool_resources: A set of resources that are used by the assistant's tools. The resources are
              specific to the type of tool. For example, the `code_interpreter` tool requires
              a list of file IDs, while the `file_search` tool requires a list of vector store
              IDs.

          tools: A list of tool enabled on the assistant. There can be a maximum of 128 tools per
              assistant. Tools can be of types `code_interpreter`, `file_search`, or
              `function`.

          top_p: An alternative to sampling with temperature, called nucleus sampling, where the
              model considers the results of the tokens with top_p probability mass. So 0.1
              means only the tokens comprising the top 10% probability mass are considered.

              We generally recommend altering this or temperature but not both.

          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        r`   rO   rP   ra   r2   r3   r4   rM   r5   r6   r7   r8   r9   r:   r;   rS   rT   )rd   rX   r   r   AssistantUpdateParamsr   r   r*   r^   r2   r3   r4   rM   r5   r6   r7   r8   r9   r:   r;   r<   r=   r>   r?   r+   r+   r,   update   s:        zAssistants.updateafterbeforelimitorderr<   r=   r>   r?   str | NotGivenint | NotGiven#Literal[('asc', 'desc')] | NotGivenzSyncCursorPage[Assistant]	rr   rs   rt   ru   r<   r=   r>   r?   r'   c          	      C  sF   ddi|pi }| j dtt t||||t||||dtjdtdS a  Returns a list of assistants.

        Args:
          after: A cursor for use in pagination.

        `after` is an object ID that defines your place
              in the list. For instance, if you make a list request and receive 100 objects,
              ending with obj_foo, your subsequent call can include after=obj_foo in order to
              fetch the next page of the list.

          before: A cursor for use in pagination. `before` is an object ID that defines your place
              in the list. For instance, if you make a list request and receive 100 objects,
              starting with obj_foo, your subsequent call can include before=obj_foo in order
              to fetch the previous page of the list.

          limit: A limit on the number of objects to be returned. Limit can range between 1 and
              100, and the default is 20.

          order: Sort order by the `created_at` timestamp of the objects. `asc` for ascending
              order and `desc` for descending order.

          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        rO   rP   rQ   )rr   rs   rt   ru   )r<   r=   r>   r?   query)pagerV   rM   )_get_api_listr   r   r   r   r   AssistantListParams	r*   rr   rs   rt   ru   r<   r=   r>   r?   r+   r+   r,   list  s(    +zAssistants.listr   c                C  sD   |st d|ddi|pi }| jd| t||||dtdS )H  
        Delete an assistant.

        Args:
          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        r`   rO   rP   ra   rS   rb   rd   _deleter   r   rf   r+   r+   r,   delete  s       zAssistants.delete__name__
__module____qualname__r   r-   r0   r	   r[   rg   rp   r   r   r+   r+   r+   r,   r#   "   sp   	0 &*2 ."Gc                   @  sF  e Zd ZeddddZeddddZeeeeeeeeeeddded	d
ddddddddddddddddddZdddeddddddddddZeeeeeeeeeeedddedddddd ddddd!dddddddd"d#d$Z	eeeeddded%d&d&d'd(ddddd)d*	d+d,Z
dddeddddddd-dd.d/ZdS )0r$   AsyncAssistantsWithRawResponser&   c                 C  s   t | S r(   )r   r)   r+   r+   r,   r-     s    z!AsyncAssistants.with_raw_response$AsyncAssistantsWithStreamingResponsec                 C  s   t | S r/   )r   r)   r+   r+   r,   r0     s    z'AsyncAssistants.with_streaming_responseNr1   r@   rA   rB   rC   rD   rE   rF   rG   rH   rI   rJ   rK   r   rL   c                  sZ   ddi|pi }| j dt|||||||||	|
|dtjI dH t||||dtdI dH S )rN   rO   rP   rQ   rR   NrS   rT   )rX   r   r   rY   r   r   rZ   r+   r+   r,   r[     s4    d   zAsyncAssistants.createrS   r\   r]   c                  sJ   |st d|ddi|pi }| jd| t||||dtdI dH S )r_   r`   rO   rP   ra   rS   rb   Nrc   rf   r+   r+   r,   rg   t  s       zAsyncAssistants.retrieverh   ri   rj   rk   c                  sr   |st d|ddi|pi }| jd| t||||||||	|
||dtjI dH t||||dtdI dH S )	rl   r`   rO   rP   ra   rm   NrS   rT   )rd   rX   r   r   rn   r   r   ro   r+   r+   r,   rp     s:        zAsyncAssistants.updaterq   rv   rw   rx   z5AsyncPaginator[Assistant, AsyncCursorPage[Assistant]]ry   c          	      C  sF   ddi|pi }| j dtt t||||t||||dtjdtdS rz   )r}   r   r   r   r   r   r~   r   r+   r+   r,   r   @  s(    +zAsyncAssistants.listr   c                  sJ   |st d|ddi|pi }| jd| t||||dtdI dH S )r   r`   rO   rP   ra   rS   rb   Nr   rf   r+   r+   r,   r     s       zAsyncAssistants.deleter   r+   r+   r+   r,   r$     sp   	0 &*2 ."Gc                   @  s   e Zd ZdddddZdS )r%   r#   None
assistantsr'   c                 C  sP   || _ t|j| _t|j| _t|j| _t|j| _t|j| _d S N)_assistantsr   Zto_raw_response_wrapperr[   rg   rp   r   r   r*   r   r+   r+   r,   __init__  s     z"AssistantsWithRawResponse.__init__Nr   r   r   r   r+   r+   r+   r,   r%     s   r%   c                   @  s   e Zd ZdddddZdS )r   r$   r   r   c                 C  sP   || _ t|j| _t|j| _t|j| _t|j| _t|j| _d S r   )r   r   Zasync_to_raw_response_wrapperr[   rg   rp   r   r   r   r+   r+   r,   r     s     z'AsyncAssistantsWithRawResponse.__init__Nr   r+   r+   r+   r,   r     s   r   c                   @  s   e Zd ZdddddZdS )r.   r#   r   r   c                 C  sF   || _ t|j| _t|j| _t|j| _t|j| _t|j| _d S r   )r   r   r[   rg   rp   r   r   r   r+   r+   r,   r     s     z(AssistantsWithStreamingResponse.__init__Nr   r+   r+   r+   r,   r.     s   r.   c                   @  s   e Zd ZdddddZdS )r   r$   r   r   c                 C  sF   || _ t|j| _t|j| _t|j| _t|j| _t|j| _d S r   )r   r   r[   rg   rp   r   r   r   r+   r+   r,   r     s     z-AsyncAssistantsWithStreamingResponse.__init__Nr   r+   r+   r+   r,   r     s   r   );
__future__r   typingr   r   r   Ztyping_extensionsr   Zhttpx r   _typesr	   r
   r   r   r   _utilsr   r   _compatr   Z	_resourcer   r   	_responser   r   Z
paginationr   r   Z
types.betar   r   r   Z_base_clientr   r   Ztypes.beta.assistantr   Ztypes.shared.chat_modelr   Ztypes.beta.assistant_deletedr   Ztypes.shared_params.metadatar   Ztypes.shared.reasoning_effortr    Ztypes.beta.assistant_tool_paramr!   Z1types.beta.assistant_response_format_option_paramr"   __all__r#   r$   r%   r   r.   r   r+   r+   r+   r,   <module>   s@      D   D