Skip to content

Commit

Permalink
Fix docstrings
Browse files Browse the repository at this point in the history
Change-Id: I6020df4e862a4f1d58462a4cd70876a8448293cf
  • Loading branch information
mayureshagashe2105 committed May 27, 2024
1 parent f48cedc commit a1c8c72
Show file tree
Hide file tree
Showing 2 changed files with 10 additions and 9 deletions.
7 changes: 1 addition & 6 deletions google/generativeai/caching.py
Original file line number Diff line number Diff line change
Expand Up @@ -210,12 +210,7 @@ def list(
yield cls._decode_cached_content(cached_content)

def delete(self, client: glm.CachedServiceClient | None = None) -> None:
"""Deletes `CachedContent` resource.
Args:
name: The resource name referring to the cached content.
Format: cachedContents/{id}.
"""
"""Deletes `CachedContent` resource."""
if client is None:
client = get_default_cache_client()

Expand Down
12 changes: 9 additions & 3 deletions tests/test_generative_models.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
import collections
from collections.abc import Iterable
import copy
import datetime
import pathlib
from typing import Any
import textwrap
Expand Down Expand Up @@ -112,7 +113,6 @@ def setUp(self):
client_lib._client_manager.clients["generative"] = self.client
client_lib._client_manager.clients["cache"] = self.client


def test_hello(self):
# Generate text from text prompt
model = generative_models.GenerativeModel(model_name="gemini-pro")
Expand Down Expand Up @@ -338,7 +338,13 @@ def test_stream_prompt_feedback_not_blocked(self):
dict(testcase_name="test_cached_content_as_id", cached_content="test-cached-content"),
dict(
testcase_name="test_cached_content_as_CachedContent_object",
cached_content=caching.CachedContent.get(name="cachedContents/test-cached-content"),
cached_content=caching.CachedContent(
name="cachedContents/test-cached-content",
model="models/gemini-1.0-pro-001",
create_time=datetime.datetime.now(),
update_time=datetime.datetime.now(),
expire_time=datetime.datetime.now(),
),
),
],
)
Expand Down Expand Up @@ -1289,7 +1295,7 @@ def test_repr_for_model_created_from_cahced_content(self):
cached_content="test-cached-content"
)
result = repr(model)
self.assertIn("cached_content=cachedContent/test-cached-content", result)
self.assertIn("cached_content=cachedContents/test-cached-content", result)
self.assertIn("model_name='models/gemini-1.0-pro-001'", result)

def test_count_tokens_called_with_request_options(self):
Expand Down

0 comments on commit a1c8c72

Please sign in to comment.