Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 6 additions & 10 deletions libs/partners/fireworks/langchain_fireworks/chat_models.py
Original file line number Diff line number Diff line change
Expand Up @@ -287,9 +287,7 @@ class ChatFireworks(BaseChatModel):
```python
from langchain_fireworks.chat_models import ChatFireworks

fireworks = ChatFireworks(
model_name="accounts/fireworks/models/llama-v3p1-8b-instruct"
)
fireworks = ChatFireworks(model_name="fireworks/llama-v3p3-70b-instruct")
```
"""

Expand Down Expand Up @@ -799,7 +797,7 @@ class AnswerWithJustification(BaseModel):


model = ChatFireworks(
model="accounts/fireworks/models/firefunction-v1",
model="fireworks/llama-v3p3-70b-instruct",
temperature=0,
)
structured_model = model.with_structured_output(AnswerWithJustification)
Expand Down Expand Up @@ -829,7 +827,7 @@ class AnswerWithJustification(BaseModel):


model = ChatFireworks(
model="accounts/fireworks/models/firefunction-v1",
model="fireworks/llama-v3p3-70b-instruct",
temperature=0,
)
structured_model = model.with_structured_output(
Expand Down Expand Up @@ -864,7 +862,7 @@ class AnswerWithJustification(TypedDict):


model = ChatFireworks(
model="accounts/fireworks/models/firefunction-v1",
model="fireworks/llama-v3p3-70b-instruct",
temperature=0,
)
structured_model = model.with_structured_output(AnswerWithJustification)
Expand Down Expand Up @@ -900,7 +898,7 @@ class AnswerWithJustification(TypedDict):
}

model = ChatFireworks(
model="accounts/fireworks/models/firefunction-v1",
model="fireworks/llama-v3p3-70b-instruct",
temperature=0,
)
structured_model = model.with_structured_output(oai_schema)
Expand All @@ -926,9 +924,7 @@ class AnswerWithJustification(BaseModel):
justification: str


model = ChatFireworks(
model="accounts/fireworks/models/firefunction-v1", temperature=0
)
model = ChatFireworks(model="fireworks/llama-v3p3-70b-instruct", temperature=0)
structured_model = model.with_structured_output(
AnswerWithJustification, method="json_mode", include_raw=True
)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@

from langchain_fireworks import ChatFireworks

_MODEL = "accounts/fireworks/models/llama-v3p1-8b-instruct"
_MODEL = "fireworks/llama-v3p3-70b-instruct"


def test_tool_choice_bool() -> None:
Expand Down Expand Up @@ -142,7 +142,7 @@ def validate_joke_dict(result: Any) -> bool:

@pytest.mark.parametrize("schema_type", ["pydantic", "typeddict", "json_schema"])
def test_structured_output_json_schema(schema_type: str) -> None:
llm = ChatFireworks(model="accounts/fireworks/models/llama-v3p1-70b-instruct")
llm = ChatFireworks(model="fireworks/llama-v3p3-70b-instruct")
schema, validation_function = _get_joke_class(schema_type) # type: ignore[arg-type]
chat = llm.with_structured_output(schema, method="json_schema")

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@

from langchain_fireworks import Fireworks

_MODEL = "accounts/fireworks/models/llama-v3p1-8b-instruct"
_MODEL = "fireworks/llama-v3p3-70b-instruct"


def test_fireworks_call() -> None:
Expand Down
Loading