@@ -192,9 +192,9 @@ async def get_current_job(
192
192
stmt = (
193
193
select (Job )
194
194
.where (
195
- ((Job .delayed_until <= curtime ) | (Job .delayed_until is None ))
196
- & (Job .completed is None )
197
- & (Job .failed is None )
195
+ ((Job .delayed_until <= curtime ) | (Job .delayed_until == None ))
196
+ & (Job .completed == None )
197
+ & (Job .failed == None )
198
198
)
199
199
.order_by (Job .priority .desc (), Job .retry .desc (), Job .id )
200
200
.limit (1 )
@@ -323,8 +323,8 @@ async def repeat_url_worker():
323
323
.join (Job )
324
324
.where (
325
325
URL .url .in_ ([job .url .url for job in jobs ])
326
- & (Job .completed is None )
327
- & (Job .failed is None )
326
+ & (Job .completed == None )
327
+ & (Job .failed == None )
328
328
)
329
329
)
330
330
result = await session .scalars (stmt2 )
@@ -359,16 +359,17 @@ async def lifespan(_: FastAPI):
359
359
global engine , async_session
360
360
if engine is None :
361
361
engine = sqlalchemy .ext .asyncio .create_async_engine (
362
- environ .get ("DATABASE_URL" , "sqlite:///db.sqlite" )
362
+ environ .get ("DATABASE_URL" , "sqlite:///db.sqlite" ),
363
+ echo = True ,
363
364
)
364
365
async_session = sqlalchemy .ext .asyncio .async_sessionmaker (
365
366
engine , expire_on_commit = False
366
367
)
367
368
async with engine .begin () as conn :
368
369
await conn .run_sync (Base .metadata .create_all )
369
- workers .append (
370
- asyncio .create_task (exception_logger (url_worker (), name = "url_worker" ))
371
- )
370
+ # workers.append(
371
+ # asyncio.create_task(exception_logger(url_worker(), name="url_worker"))
372
+ # )
372
373
workers .append (
373
374
asyncio .create_task (
374
375
exception_logger (repeat_url_worker (), name = "repeat_url_worker" )
@@ -453,7 +454,7 @@ async def stats() -> Stats:
453
454
(
454
455
await session .execute (
455
456
select (Job .retry , sqlalchemy .func .count (Job .id ))
456
- .where ((Job .completed is None ) & (Job .failed is None ))
457
+ .where ((Job .completed == None ) & (Job .failed == None ))
457
458
.group_by (Job .retry )
458
459
)
459
460
).all ()
@@ -462,21 +463,21 @@ async def stats() -> Stats:
462
463
(
463
464
await session .execute (
464
465
select (Job .retry , sqlalchemy .func .count (Job .id ))
465
- .where (Job .completed is not None )
466
+ .where (Job .completed != None )
466
467
.group_by (Job .retry )
467
468
)
468
469
).all ()
469
470
)
470
471
failed = (
471
472
await session .scalar (
472
- select (sqlalchemy .func .count (Job .id )).where (Job .failed is not None )
473
+ select (sqlalchemy .func .count (Job .id )).where (Job .failed != None )
473
474
)
474
475
) or 0
475
476
batches = (await session .scalar (select (sqlalchemy .func .count (Batch .id )))) or 0
476
477
super_recently_archived_urls = (
477
478
await session .scalar (
478
479
select (sqlalchemy .func .count (URL .id ))
479
- .where (URL .last_seen is not None )
480
+ .where (URL .last_seen != None )
480
481
.where (
481
482
URL .last_seen
482
483
> datetime .datetime .now (tz = datetime .timezone .utc )
@@ -488,7 +489,7 @@ async def stats() -> Stats:
488
489
(
489
490
await session .scalar (
490
491
select (sqlalchemy .func .count (URL .id ))
491
- .where (URL .last_seen is not None )
492
+ .where (URL .last_seen != None )
492
493
.where (
493
494
URL .last_seen
494
495
> datetime .datetime .now (tz = datetime .timezone .utc )
@@ -501,7 +502,7 @@ async def stats() -> Stats:
501
502
not_recently_archived_urls = (
502
503
await session .scalar (
503
504
select (sqlalchemy .func .count (URL .id ))
504
- .where (URL .last_seen is not None )
505
+ .where (URL .last_seen != None )
505
506
.where (
506
507
URL .last_seen
507
508
< datetime .datetime .now (tz = datetime .timezone .utc )
@@ -511,20 +512,20 @@ async def stats() -> Stats:
511
512
) or 0
512
513
not_archived_urls = (
513
514
await session .scalar (
514
- select (sqlalchemy .func .count (URL .id )).where (URL .last_seen is None )
515
+ select (sqlalchemy .func .count (URL .id )).where (URL .last_seen == None )
515
516
)
516
517
) or 0
517
518
active_repeat_urls = (
518
519
await session .scalar (
519
520
select (sqlalchemy .func .count (RepeatURL .id )).where (
520
- RepeatURL .active_since is not None
521
+ RepeatURL .active_since != None
521
522
)
522
523
)
523
524
) or 0
524
525
inactive_repeat_urls = (
525
526
await session .scalar (
526
527
select (sqlalchemy .func .count (RepeatURL .id )).where (
527
- RepeatURL .active_since is None
528
+ RepeatURL .active_since == None
528
529
)
529
530
)
530
531
) or 0
@@ -676,7 +677,7 @@ def from_job(cls, job: Job, batch_ids: list[int] | None = None):
676
677
retry = job .retry ,
677
678
failed = job .failed ,
678
679
batches = batch_ids
679
- if batch_ids is not None
680
+ if batch_ids != None
680
681
else [batch .id for batch in job .batches ],
681
682
)
682
683
@@ -867,16 +868,16 @@ def apply_job_filtering(
867
868
# If all 4 are given, we can take a shortcut and not apply anything
868
869
if query_params ["not_started" ]:
869
870
in_statement = in_statement .where (
870
- (Job .completed is None )
871
- & (Job .failed is None )
872
- & (Job .delayed_until is None )
871
+ (Job .completed == None )
872
+ & (Job .failed == None )
873
+ & (Job .delayed_until == None )
873
874
)
874
875
if query_params ["completed" ]:
875
- in_statement = in_statement .where (Job .completed is not None )
876
+ in_statement = in_statement .where (Job .completed != None )
876
877
if query_params ["delayed" ]:
877
- in_statement = in_statement .where (Job .delayed_until is not None )
878
+ in_statement = in_statement .where (Job .delayed_until != None )
878
879
if query_params ["failed" ]:
879
- in_statement = in_statement .where (Job .failed is not None )
880
+ in_statement = in_statement .where (Job .failed != None )
880
881
retry_param_count = [
881
882
query_params ["retries_less_than" ],
882
883
query_params ["retries_greater_than" ],
0 commit comments