Compare commits
12 Commits
Author | SHA1 | Date | |
---|---|---|---|
![]() |
0c8192222f | ||
![]() |
9c9602363c | ||
![]() |
27bad2575e | ||
![]() |
3a7497c852 | ||
![]() |
77594bd0ea | ||
![]() |
e4ccba7aee | ||
![]() |
09b2d92466 | ||
![]() |
cb087acbeb | ||
![]() |
7129bfb513 | ||
![]() |
b41b1f364a | ||
![]() |
9ebf16c140 | ||
![]() |
6d0fae36e6 |
@ -1,6 +1,6 @@
|
||||
[tool.poetry]
|
||||
name = "reflex"
|
||||
version = "0.6.6dev1"
|
||||
version = "0.6.6"
|
||||
description = "Web apps in pure Python."
|
||||
license = "Apache-2.0"
|
||||
authors = [
|
||||
|
@ -454,6 +454,10 @@ export const connect = async (
|
||||
queueEvents(update.events, socket);
|
||||
}
|
||||
});
|
||||
socket.current.on("reload", async (event) => {
|
||||
event_processing = false;
|
||||
queueEvents([...initialEvents(), JSON5.parse(event)], socket);
|
||||
});
|
||||
|
||||
document.addEventListener("visibilitychange", checkVisibility);
|
||||
};
|
||||
@ -486,23 +490,30 @@ export const uploadFiles = async (
|
||||
return false;
|
||||
}
|
||||
|
||||
// Track how many partial updates have been processed for this upload.
|
||||
let resp_idx = 0;
|
||||
const eventHandler = (progressEvent) => {
|
||||
// handle any delta / event streamed from the upload event handler
|
||||
const event_callbacks = socket._callbacks.$event;
|
||||
// Whenever called, responseText will contain the entire response so far.
|
||||
const chunks = progressEvent.event.target.responseText.trim().split("\n");
|
||||
// So only process _new_ chunks beyond resp_idx.
|
||||
chunks.slice(resp_idx).map((chunk) => {
|
||||
try {
|
||||
socket._callbacks.$event.map((f) => {
|
||||
f(chunk);
|
||||
});
|
||||
resp_idx += 1;
|
||||
} catch (e) {
|
||||
if (progressEvent.progress === 1) {
|
||||
// Chunk may be incomplete, so only report errors when full response is available.
|
||||
console.log("Error parsing chunk", chunk, e);
|
||||
}
|
||||
return;
|
||||
}
|
||||
event_callbacks.map((f, ix) => {
|
||||
f(chunk)
|
||||
.then(() => {
|
||||
if (ix === event_callbacks.length - 1) {
|
||||
// Mark this chunk as processed.
|
||||
resp_idx += 1;
|
||||
}
|
||||
})
|
||||
.catch((e) => {
|
||||
if (progressEvent.progress === 1) {
|
||||
// Chunk may be incomplete, so only report errors when full response is available.
|
||||
console.log("Error parsing chunk", chunk, e);
|
||||
}
|
||||
return;
|
||||
});
|
||||
});
|
||||
});
|
||||
};
|
||||
|
||||
@ -707,7 +718,7 @@ export const useEventLoop = (
|
||||
const combined_name = events.map((e) => e.name).join("+++");
|
||||
if (event_actions?.temporal) {
|
||||
if (!socket.current || !socket.current.connected) {
|
||||
return; // don't queue when the backend is not connected
|
||||
return; // don't queue when the backend is not connected
|
||||
}
|
||||
}
|
||||
if (event_actions?.throttle) {
|
||||
@ -848,7 +859,7 @@ export const useEventLoop = (
|
||||
if (router.components[router.pathname].error) {
|
||||
delete router.components[router.pathname].error;
|
||||
}
|
||||
}
|
||||
};
|
||||
router.events.on("routeChangeStart", change_start);
|
||||
router.events.on("routeChangeComplete", change_complete);
|
||||
router.events.on("routeChangeError", change_error);
|
||||
|
@ -73,6 +73,7 @@ from reflex.event import (
|
||||
EventSpec,
|
||||
EventType,
|
||||
IndividualEventType,
|
||||
get_hydrate_event,
|
||||
window_alert,
|
||||
)
|
||||
from reflex.model import Model, get_db_status
|
||||
@ -1259,6 +1260,21 @@ async def process(
|
||||
)
|
||||
# Get the state for the session exclusively.
|
||||
async with app.state_manager.modify_state(event.substate_token) as state:
|
||||
# When this is a brand new instance of the state, signal the
|
||||
# frontend to reload before processing it.
|
||||
if (
|
||||
not state.router_data
|
||||
and event.name != get_hydrate_event(state)
|
||||
and app.event_namespace is not None
|
||||
):
|
||||
await asyncio.create_task(
|
||||
app.event_namespace.emit(
|
||||
"reload",
|
||||
data=format.json_dumps(event),
|
||||
to=sid,
|
||||
)
|
||||
)
|
||||
return
|
||||
# re-assign only when the value is different
|
||||
if state.router_data != router_data:
|
||||
# assignment will recurse into substates and force recalculation of
|
||||
@ -1462,10 +1478,10 @@ class EventNamespace(AsyncNamespace):
|
||||
app: App
|
||||
|
||||
# Keep a mapping between socket ID and client token.
|
||||
token_to_sid: dict[str, str] = {}
|
||||
token_to_sid: dict[str, str]
|
||||
|
||||
# Keep a mapping between client token and socket ID.
|
||||
sid_to_token: dict[str, str] = {}
|
||||
sid_to_token: dict[str, str]
|
||||
|
||||
def __init__(self, namespace: str, app: App):
|
||||
"""Initialize the event namespace.
|
||||
@ -1475,6 +1491,8 @@ class EventNamespace(AsyncNamespace):
|
||||
app: The application object.
|
||||
"""
|
||||
super().__init__(namespace)
|
||||
self.token_to_sid = {}
|
||||
self.sid_to_token = {}
|
||||
self.app = app
|
||||
|
||||
def on_connect(self, sid, environ):
|
||||
|
@ -293,13 +293,15 @@ class Upload(MemoizationLeaf):
|
||||
format.to_camel_case(key): value for key, value in upload_props.items()
|
||||
}
|
||||
|
||||
use_dropzone_arguments = {
|
||||
"onDrop": event_var,
|
||||
**upload_props,
|
||||
}
|
||||
use_dropzone_arguments = Var.create(
|
||||
{
|
||||
"onDrop": event_var,
|
||||
**upload_props,
|
||||
}
|
||||
)
|
||||
|
||||
left_side = f"const {{getRootProps: {root_props_unique_name}, getInputProps: {input_props_unique_name}}} "
|
||||
right_side = f"useDropzone({str(Var.create(use_dropzone_arguments))})"
|
||||
right_side = f"useDropzone({str(use_dropzone_arguments)})"
|
||||
|
||||
var_data = VarData.merge(
|
||||
VarData(
|
||||
@ -307,6 +309,7 @@ class Upload(MemoizationLeaf):
|
||||
hooks={Hooks.EVENTS: None},
|
||||
),
|
||||
event_var._get_all_var_data(),
|
||||
use_dropzone_arguments._get_all_var_data(),
|
||||
VarData(
|
||||
hooks={
|
||||
callback_str: None,
|
||||
|
@ -255,7 +255,7 @@ const extractPoints = (points) => {
|
||||
|
||||
def _render(self):
|
||||
tag = super()._render()
|
||||
figure = self.data.to(dict)
|
||||
figure = self.data.to(dict) if self.data is not None else Var.create({})
|
||||
merge_dicts = [] # Data will be merged and spread from these dict Vars
|
||||
if self.layout is not None:
|
||||
# Why is this not a literal dict? Great question... it didn't work
|
||||
|
@ -3,7 +3,6 @@
|
||||
from typing import Dict, Literal
|
||||
|
||||
from reflex.components.component import Component, MemoizationLeaf, NoSSRComponent
|
||||
from reflex.utils import console
|
||||
|
||||
|
||||
class Recharts(Component):
|
||||
@ -11,19 +10,8 @@ class Recharts(Component):
|
||||
|
||||
library = "recharts@2.13.0"
|
||||
|
||||
def render(self) -> Dict:
|
||||
"""Render the tag.
|
||||
|
||||
Returns:
|
||||
The rendered tag.
|
||||
"""
|
||||
tag = super().render()
|
||||
if any(p.startswith("css") for p in tag["props"]):
|
||||
console.warn(
|
||||
f"CSS props do not work for {self.__class__.__name__}. Consult docs to style it with its own prop."
|
||||
)
|
||||
tag["props"] = [p for p in tag["props"] if not p.startswith("css")]
|
||||
return tag
|
||||
def _get_style(self) -> Dict:
|
||||
return {"wrapperStyle": self.style}
|
||||
|
||||
|
||||
class RechartsCharts(NoSSRComponent, MemoizationLeaf):
|
||||
|
@ -11,7 +11,6 @@ from reflex.style import Style
|
||||
from reflex.vars.base import Var
|
||||
|
||||
class Recharts(Component):
|
||||
def render(self) -> Dict: ...
|
||||
@overload
|
||||
@classmethod
|
||||
def create( # type: ignore
|
||||
|
@ -404,7 +404,7 @@ def logoutv2(
|
||||
|
||||
hosting.log_out_on_browser()
|
||||
console.debug("Deleting access token from config locally")
|
||||
hosting.delete_token_from_config(include_invitation_code=True)
|
||||
hosting.delete_token_from_config()
|
||||
|
||||
|
||||
db_cli = typer.Typer()
|
||||
|
@ -1748,7 +1748,11 @@ class BaseState(Base, ABC, extra=pydantic.Extra.allow):
|
||||
if value is None:
|
||||
continue
|
||||
hinted_args = value_inside_optional(hinted_args)
|
||||
if isinstance(value, dict) and inspect.isclass(hinted_args):
|
||||
if (
|
||||
isinstance(value, dict)
|
||||
and inspect.isclass(hinted_args)
|
||||
and not types.is_generic_alias(hinted_args) # py3.9-py3.10
|
||||
):
|
||||
if issubclass(hinted_args, Model):
|
||||
# Remove non-fields from the payload
|
||||
payload[arg] = hinted_args(
|
||||
@ -1759,7 +1763,7 @@ class BaseState(Base, ABC, extra=pydantic.Extra.allow):
|
||||
}
|
||||
)
|
||||
elif dataclasses.is_dataclass(hinted_args) or issubclass(
|
||||
hinted_args, Base
|
||||
hinted_args, (Base, BaseModelV1, BaseModelV2)
|
||||
):
|
||||
payload[arg] = hinted_args(**value)
|
||||
if isinstance(value, list) and (hinted_args is set or hinted_args is Set):
|
||||
@ -1955,6 +1959,9 @@ class BaseState(Base, ABC, extra=pydantic.Extra.allow):
|
||||
if var in self.base_vars or var in self._backend_vars:
|
||||
self._was_touched = True
|
||||
break
|
||||
if var == constants.ROUTER_DATA and self.parent_state is None:
|
||||
self._was_touched = True
|
||||
break
|
||||
|
||||
def _get_was_touched(self) -> bool:
|
||||
"""Check current dirty_vars and flag to determine if state instance was modified.
|
||||
|
@ -10,6 +10,13 @@ from selenium.webdriver import Firefox
|
||||
from selenium.webdriver.common.by import By
|
||||
from selenium.webdriver.remote.webdriver import WebDriver
|
||||
|
||||
from reflex.state import (
|
||||
State,
|
||||
StateManagerDisk,
|
||||
StateManagerMemory,
|
||||
StateManagerRedis,
|
||||
_substate_key,
|
||||
)
|
||||
from reflex.testing import AppHarness
|
||||
|
||||
from . import utils
|
||||
@ -74,7 +81,7 @@ def ClientSide():
|
||||
return rx.fragment(
|
||||
rx.input(
|
||||
value=ClientSideState.router.session.client_token,
|
||||
is_read_only=True,
|
||||
read_only=True,
|
||||
id="token",
|
||||
),
|
||||
rx.input(
|
||||
@ -604,6 +611,110 @@ async def test_client_side_state(
|
||||
assert s2.text == "s2 value"
|
||||
assert s3.text == "s3 value"
|
||||
|
||||
# Simulate state expiration
|
||||
if isinstance(client_side.state_manager, StateManagerRedis):
|
||||
await client_side.state_manager.redis.delete(
|
||||
_substate_key(token, State.get_full_name())
|
||||
)
|
||||
await client_side.state_manager.redis.delete(_substate_key(token, state_name))
|
||||
await client_side.state_manager.redis.delete(
|
||||
_substate_key(token, sub_state_name)
|
||||
)
|
||||
await client_side.state_manager.redis.delete(
|
||||
_substate_key(token, sub_sub_state_name)
|
||||
)
|
||||
elif isinstance(client_side.state_manager, (StateManagerMemory, StateManagerDisk)):
|
||||
del client_side.state_manager.states[token]
|
||||
if isinstance(client_side.state_manager, StateManagerDisk):
|
||||
client_side.state_manager.token_expiration = 0
|
||||
client_side.state_manager._purge_expired_states()
|
||||
|
||||
# Ensure the state is gone (not hydrated)
|
||||
async def poll_for_not_hydrated():
|
||||
state = await client_side.get_state(_substate_key(token or "", state_name))
|
||||
return not state.is_hydrated
|
||||
|
||||
assert await AppHarness._poll_for_async(poll_for_not_hydrated)
|
||||
|
||||
# Trigger event to get a new instance of the state since the old was expired.
|
||||
state_var_input = driver.find_element(By.ID, "state_var")
|
||||
state_var_input.send_keys("re-triggering")
|
||||
|
||||
# get new references to all cookie and local storage elements (again)
|
||||
c1 = driver.find_element(By.ID, "c1")
|
||||
c2 = driver.find_element(By.ID, "c2")
|
||||
c3 = driver.find_element(By.ID, "c3")
|
||||
c4 = driver.find_element(By.ID, "c4")
|
||||
c5 = driver.find_element(By.ID, "c5")
|
||||
c6 = driver.find_element(By.ID, "c6")
|
||||
c7 = driver.find_element(By.ID, "c7")
|
||||
l1 = driver.find_element(By.ID, "l1")
|
||||
l2 = driver.find_element(By.ID, "l2")
|
||||
l3 = driver.find_element(By.ID, "l3")
|
||||
l4 = driver.find_element(By.ID, "l4")
|
||||
s1 = driver.find_element(By.ID, "s1")
|
||||
s2 = driver.find_element(By.ID, "s2")
|
||||
s3 = driver.find_element(By.ID, "s3")
|
||||
c1s = driver.find_element(By.ID, "c1s")
|
||||
l1s = driver.find_element(By.ID, "l1s")
|
||||
s1s = driver.find_element(By.ID, "s1s")
|
||||
|
||||
assert c1.text == "c1 value"
|
||||
assert c2.text == "c2 value"
|
||||
assert c3.text == "" # temporary cookie expired after reset state!
|
||||
assert c4.text == "c4 value"
|
||||
assert c5.text == "c5 value"
|
||||
assert c6.text == "c6 value"
|
||||
assert c7.text == "c7 value"
|
||||
assert l1.text == "l1 value"
|
||||
assert l2.text == "l2 value"
|
||||
assert l3.text == "l3 value"
|
||||
assert l4.text == "l4 value"
|
||||
assert s1.text == "s1 value"
|
||||
assert s2.text == "s2 value"
|
||||
assert s3.text == "s3 value"
|
||||
assert c1s.text == "c1s value"
|
||||
assert l1s.text == "l1s value"
|
||||
assert s1s.text == "s1s value"
|
||||
|
||||
# Get the backend state and ensure the values are still set
|
||||
async def get_sub_state():
|
||||
root_state = await client_side.get_state(
|
||||
_substate_key(token or "", sub_state_name)
|
||||
)
|
||||
state = root_state.substates[client_side.get_state_name("_client_side_state")]
|
||||
sub_state = state.substates[
|
||||
client_side.get_state_name("_client_side_sub_state")
|
||||
]
|
||||
return sub_state
|
||||
|
||||
async def poll_for_c1_set():
|
||||
sub_state = await get_sub_state()
|
||||
return sub_state.c1 == "c1 value"
|
||||
|
||||
assert await AppHarness._poll_for_async(poll_for_c1_set)
|
||||
sub_state = await get_sub_state()
|
||||
assert sub_state.c1 == "c1 value"
|
||||
assert sub_state.c2 == "c2 value"
|
||||
assert sub_state.c3 == ""
|
||||
assert sub_state.c4 == "c4 value"
|
||||
assert sub_state.c5 == "c5 value"
|
||||
assert sub_state.c6 == "c6 value"
|
||||
assert sub_state.c7 == "c7 value"
|
||||
assert sub_state.l1 == "l1 value"
|
||||
assert sub_state.l2 == "l2 value"
|
||||
assert sub_state.l3 == "l3 value"
|
||||
assert sub_state.l4 == "l4 value"
|
||||
assert sub_state.s1 == "s1 value"
|
||||
assert sub_state.s2 == "s2 value"
|
||||
assert sub_state.s3 == "s3 value"
|
||||
sub_sub_state = sub_state.substates[
|
||||
client_side.get_state_name("_client_side_sub_sub_state")
|
||||
]
|
||||
assert sub_sub_state.c1s == "c1s value"
|
||||
assert sub_sub_state.l1s == "l1s value"
|
||||
assert sub_sub_state.s1s == "s1s value"
|
||||
|
||||
# clear the cookie jar and local storage, ensure state reset to default
|
||||
driver.delete_all_cookies()
|
||||
local_storage.clear()
|
||||
|
@ -19,10 +19,14 @@ def UploadFile():
|
||||
|
||||
import reflex as rx
|
||||
|
||||
LARGE_DATA = "DUMMY" * 1024 * 512
|
||||
|
||||
class UploadState(rx.State):
|
||||
_file_data: Dict[str, str] = {}
|
||||
event_order: List[str] = []
|
||||
progress_dicts: List[dict] = []
|
||||
disabled: bool = False
|
||||
large_data: str = ""
|
||||
|
||||
async def handle_upload(self, files: List[rx.UploadFile]):
|
||||
for file in files:
|
||||
@ -33,6 +37,7 @@ def UploadFile():
|
||||
for file in files:
|
||||
upload_data = await file.read()
|
||||
self._file_data[file.filename or ""] = upload_data.decode("utf-8")
|
||||
self.large_data = LARGE_DATA
|
||||
yield UploadState.chain_event
|
||||
|
||||
def upload_progress(self, progress):
|
||||
@ -41,13 +46,15 @@ def UploadFile():
|
||||
self.progress_dicts.append(progress)
|
||||
|
||||
def chain_event(self):
|
||||
assert self.large_data == LARGE_DATA
|
||||
self.large_data = ""
|
||||
self.event_order.append("chain_event")
|
||||
|
||||
def index():
|
||||
return rx.vstack(
|
||||
rx.input(
|
||||
value=UploadState.router.session.client_token,
|
||||
is_read_only=True,
|
||||
read_only=True,
|
||||
id="token",
|
||||
),
|
||||
rx.heading("Default Upload"),
|
||||
@ -56,6 +63,7 @@ def UploadFile():
|
||||
rx.button("Select File"),
|
||||
rx.text("Drag and drop files here or click to select files"),
|
||||
),
|
||||
disabled=UploadState.disabled,
|
||||
),
|
||||
rx.button(
|
||||
"Upload",
|
||||
|
@ -1007,8 +1007,9 @@ async def test_dynamic_route_var_route_change_completed_on_load(
|
||||
substate_token = _substate_key(token, DynamicState)
|
||||
sid = "mock_sid"
|
||||
client_ip = "127.0.0.1"
|
||||
state = await app.state_manager.get_state(substate_token)
|
||||
assert state.dynamic == ""
|
||||
async with app.state_manager.modify_state(substate_token) as state:
|
||||
state.router_data = {"simulate": "hydrated"}
|
||||
assert state.dynamic == ""
|
||||
exp_vals = ["foo", "foobar", "baz"]
|
||||
|
||||
def _event(name, val, **kwargs):
|
||||
@ -1180,6 +1181,7 @@ async def test_process_events(mocker, token: str):
|
||||
"ip": "127.0.0.1",
|
||||
}
|
||||
app = App(state=GenState)
|
||||
|
||||
mocker.patch.object(app, "_postprocess", AsyncMock())
|
||||
event = Event(
|
||||
token=token,
|
||||
@ -1187,6 +1189,8 @@ async def test_process_events(mocker, token: str):
|
||||
payload={"c": 5},
|
||||
router_data=router_data,
|
||||
)
|
||||
async with app.state_manager.modify_state(event.substate_token) as state:
|
||||
state.router_data = {"simulate": "hydrated"}
|
||||
|
||||
async for _update in process(app, event, "mock_sid", {}, "127.0.0.1"):
|
||||
pass
|
||||
|
@ -10,7 +10,17 @@ import os
|
||||
import sys
|
||||
import threading
|
||||
from textwrap import dedent
|
||||
from typing import Any, AsyncGenerator, Callable, Dict, List, Optional, Union
|
||||
from typing import (
|
||||
Any,
|
||||
AsyncGenerator,
|
||||
Callable,
|
||||
Dict,
|
||||
List,
|
||||
Optional,
|
||||
Set,
|
||||
Tuple,
|
||||
Union,
|
||||
)
|
||||
from unittest.mock import AsyncMock, Mock
|
||||
|
||||
import pytest
|
||||
@ -1828,12 +1838,11 @@ async def test_state_manager_lock_expire_contend(
|
||||
|
||||
|
||||
@pytest.fixture(scope="function")
|
||||
def mock_app(monkeypatch, state_manager: StateManager) -> rx.App:
|
||||
"""Mock app fixture.
|
||||
def mock_app_simple(monkeypatch) -> rx.App:
|
||||
"""Simple Mock app fixture.
|
||||
|
||||
Args:
|
||||
monkeypatch: Pytest monkeypatch object.
|
||||
state_manager: A state manager.
|
||||
|
||||
Returns:
|
||||
The app, after mocking out prerequisites.get_app()
|
||||
@ -1844,7 +1853,6 @@ def mock_app(monkeypatch, state_manager: StateManager) -> rx.App:
|
||||
|
||||
setattr(app_module, CompileVars.APP, app)
|
||||
app.state = TestState
|
||||
app._state_manager = state_manager
|
||||
app.event_namespace.emit = AsyncMock() # type: ignore
|
||||
|
||||
def _mock_get_app(*args, **kwargs):
|
||||
@ -1854,6 +1862,21 @@ def mock_app(monkeypatch, state_manager: StateManager) -> rx.App:
|
||||
return app
|
||||
|
||||
|
||||
@pytest.fixture(scope="function")
|
||||
def mock_app(mock_app_simple: rx.App, state_manager: StateManager) -> rx.App:
|
||||
"""Mock app fixture.
|
||||
|
||||
Args:
|
||||
mock_app_simple: A simple mock app.
|
||||
state_manager: A state manager.
|
||||
|
||||
Returns:
|
||||
The app, after mocking out prerequisites.get_app()
|
||||
"""
|
||||
mock_app_simple._state_manager = state_manager
|
||||
return mock_app_simple
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_state_proxy(grandchild_state: GrandchildState, mock_app: rx.App):
|
||||
"""Test that the state proxy works.
|
||||
@ -1959,6 +1982,10 @@ class BackgroundTaskState(BaseState):
|
||||
order: List[str] = []
|
||||
dict_list: Dict[str, List[int]] = {"foo": [1, 2, 3]}
|
||||
|
||||
def __init__(self, **kwargs): # noqa: D107
|
||||
super().__init__(**kwargs)
|
||||
self.router_data = {"simulate": "hydrate"}
|
||||
|
||||
@rx.var
|
||||
def computed_order(self) -> List[str]:
|
||||
"""Get the order as a computed var.
|
||||
@ -2709,7 +2736,7 @@ def test_set_base_field_via_setter():
|
||||
assert "c2" in bfss.dirty_vars
|
||||
|
||||
|
||||
def exp_is_hydrated(state: State, is_hydrated: bool = True) -> Dict[str, Any]:
|
||||
def exp_is_hydrated(state: BaseState, is_hydrated: bool = True) -> Dict[str, Any]:
|
||||
"""Expected IS_HYDRATED delta that would be emitted by HydrateMiddleware.
|
||||
|
||||
Args:
|
||||
@ -2788,7 +2815,8 @@ async def test_preprocess(app_module_mock, token, test_state, expected, mocker):
|
||||
app = app_module_mock.app = App(
|
||||
state=State, load_events={"index": [test_state.test_handler]}
|
||||
)
|
||||
state = State()
|
||||
async with app.state_manager.modify_state(_substate_key(token, State)) as state:
|
||||
state.router_data = {"simulate": "hydrate"}
|
||||
|
||||
updates = []
|
||||
async for update in rx.app.process(
|
||||
@ -2835,7 +2863,8 @@ async def test_preprocess_multiple_load_events(app_module_mock, token, mocker):
|
||||
state=State,
|
||||
load_events={"index": [OnLoadState.test_handler, OnLoadState.test_handler]},
|
||||
)
|
||||
state = State()
|
||||
async with app.state_manager.modify_state(_substate_key(token, State)) as state:
|
||||
state.router_data = {"simulate": "hydrate"}
|
||||
|
||||
updates = []
|
||||
async for update in rx.app.process(
|
||||
@ -3506,3 +3535,106 @@ def test_init_mixin() -> None:
|
||||
|
||||
with pytest.raises(ReflexRuntimeError):
|
||||
SubMixin()
|
||||
|
||||
|
||||
class ReflexModel(rx.Model):
|
||||
"""A model for testing."""
|
||||
|
||||
foo: str
|
||||
|
||||
|
||||
class UpcastState(rx.State):
|
||||
"""A state for testing upcasting."""
|
||||
|
||||
passed: bool = False
|
||||
|
||||
def rx_model(self, m: ReflexModel): # noqa: D102
|
||||
assert isinstance(m, ReflexModel)
|
||||
self.passed = True
|
||||
|
||||
def rx_base(self, o: Object): # noqa: D102
|
||||
assert isinstance(o, Object)
|
||||
self.passed = True
|
||||
|
||||
def rx_base_or_none(self, o: Optional[Object]): # noqa: D102
|
||||
if o is not None:
|
||||
assert isinstance(o, Object)
|
||||
self.passed = True
|
||||
|
||||
def rx_basemodelv1(self, m: ModelV1): # noqa: D102
|
||||
assert isinstance(m, ModelV1)
|
||||
self.passed = True
|
||||
|
||||
def rx_basemodelv2(self, m: ModelV2): # noqa: D102
|
||||
assert isinstance(m, ModelV2)
|
||||
self.passed = True
|
||||
|
||||
def rx_dataclass(self, dc: ModelDC): # noqa: D102
|
||||
assert isinstance(dc, ModelDC)
|
||||
self.passed = True
|
||||
|
||||
def py_set(self, s: set): # noqa: D102
|
||||
assert isinstance(s, set)
|
||||
self.passed = True
|
||||
|
||||
def py_Set(self, s: Set): # noqa: D102
|
||||
assert isinstance(s, Set)
|
||||
self.passed = True
|
||||
|
||||
def py_tuple(self, t: tuple): # noqa: D102
|
||||
assert isinstance(t, tuple)
|
||||
self.passed = True
|
||||
|
||||
def py_Tuple(self, t: Tuple): # noqa: D102
|
||||
assert isinstance(t, tuple)
|
||||
self.passed = True
|
||||
|
||||
def py_dict(self, d: dict[str, str]): # noqa: D102
|
||||
assert isinstance(d, dict)
|
||||
self.passed = True
|
||||
|
||||
def py_list(self, ls: list[str]): # noqa: D102
|
||||
assert isinstance(ls, list)
|
||||
self.passed = True
|
||||
|
||||
def py_Any(self, a: Any): # noqa: D102
|
||||
assert isinstance(a, list)
|
||||
self.passed = True
|
||||
|
||||
def py_unresolvable(self, u: "Unresolvable"): # noqa: D102, F821 # type: ignore
|
||||
assert isinstance(u, list)
|
||||
self.passed = True
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.usefixtures("mock_app_simple")
|
||||
@pytest.mark.parametrize(
|
||||
("handler", "payload"),
|
||||
[
|
||||
(UpcastState.rx_model, {"m": {"foo": "bar"}}),
|
||||
(UpcastState.rx_base, {"o": {"foo": "bar"}}),
|
||||
(UpcastState.rx_base_or_none, {"o": {"foo": "bar"}}),
|
||||
(UpcastState.rx_base_or_none, {"o": None}),
|
||||
(UpcastState.rx_basemodelv1, {"m": {"foo": "bar"}}),
|
||||
(UpcastState.rx_basemodelv2, {"m": {"foo": "bar"}}),
|
||||
(UpcastState.rx_dataclass, {"dc": {"foo": "bar"}}),
|
||||
(UpcastState.py_set, {"s": ["foo", "foo"]}),
|
||||
(UpcastState.py_Set, {"s": ["foo", "foo"]}),
|
||||
(UpcastState.py_tuple, {"t": ["foo", "foo"]}),
|
||||
(UpcastState.py_Tuple, {"t": ["foo", "foo"]}),
|
||||
(UpcastState.py_dict, {"d": {"foo": "bar"}}),
|
||||
(UpcastState.py_list, {"ls": ["foo", "foo"]}),
|
||||
(UpcastState.py_Any, {"a": ["foo"]}),
|
||||
(UpcastState.py_unresolvable, {"u": ["foo"]}),
|
||||
],
|
||||
)
|
||||
async def test_upcast_event_handler_arg(handler, payload):
|
||||
"""Test that upcast event handler args work correctly.
|
||||
|
||||
Args:
|
||||
handler: The handler to test.
|
||||
payload: The payload to test.
|
||||
"""
|
||||
state = UpcastState()
|
||||
async for update in state._process_event(handler, state, payload):
|
||||
assert update.delta == {UpcastState.get_full_name(): {"passed": True}}
|
||||
|
Loading…
Reference in New Issue
Block a user