Skip to content

Test test_timeout_in_handler is flaky #1652

@vdusek

Description

@vdusek

https://github.com/apify/crawlee-python/actions/runs/20395848580/job/58611220603?pr=1635

=========================== short test summary info ============================
FAILED tests/unit/crawlers/_basic/test_basic_crawler.py::test_timeout_in_handler[async_sleep] - assert 0 == 1
 +  where 0 = StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2025, 12, 20, 14, 41, 27, 856866, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=946305), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0).requests_finished
 +    where StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2025, 12, 20, 14, 41, 27, 856866, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=946305), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0) = <crawlee.statistics._statistics.Statistics object at 0x110733a10>.state
 +      where <crawlee.statistics._statistics.Statistics object at 0x110733a10> = <crawlee.crawlers._basic._basic_crawler.BasicCrawler object at 0x110732e40>.statistics
======================== 1 failed, 1 skipped in 16.87s =========================

https://github.com/apify/crawlee-python/actions/runs/20742769552/job/59552920657

FAILED tests/unit/crawlers/_basic/test_basic_crawler.py::test_timeout_in_handler[async_sleep] - assert 0 == 1
 +  where 0 = StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 37, 35, 531091, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=997703), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0).requests_finished
 +    where StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 37, 35, 531091, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=997703), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0) = <crawlee.statistics._statistics.Statistics object at 0x130f02c90>.state
 +      where <crawlee.statistics._statistics.Statistics object at 0x130f02c90> = <crawlee.crawlers._basic._basic_crawler.BasicCrawler object at 0x130ced850>.statistics
======================== 1 failed, 1 skipped in 22.67s =========================

https://github.com/apify/crawlee-python/actions/runs/20742882768/job/59553465201

FAILED tests/unit/crawlers/_basic/test_basic_crawler.py::test_timeout_in_handler[async_sleep] - assert 0 == 1
 +  where 0 = StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 46, 56, 154946, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=959784), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0).requests_finished
 +    where StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 46, 56, 154946, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=959784), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0) = <crawlee.statistics._statistics.Statistics object at 0x11ec4ac90>.state
 +      where <crawlee.statistics._statistics.Statistics object at 0x11ec4ac90> = <crawlee.crawlers._basic._basic_crawler.BasicCrawler object at 0x12956de10>.statistics
======================== 1 failed, 1 skipped in 21.65s =========================

Metadata

Metadata

Assignees

Labels

t-toolingIssues with this label are in the ownership of the tooling team.

Type

No type
No fields configured for issues without a type.

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions