Przeglądaj źródła

chore: bump ruff to 0.11.0 and fix linting violations (#15953)

Bowen Liang 1 miesiąc temu
rodzic
commit
9e782d4c1e

+ 1 - 1
api/core/app/apps/base_app_generator.py

@@ -151,7 +151,7 @@ class BaseAppGenerator:
 
             def gen():
                 for message in generator:
-                    if isinstance(message, (Mapping, dict)):
+                    if isinstance(message, Mapping | dict):
                         yield f"data: {json.dumps(message)}\n\n"
                     else:
                         yield f"event: {message}\n\n"

+ 1 - 2
api/core/workflow/entities/variable_pool.py

@@ -15,7 +15,6 @@ from ..enums import SystemVariableKey
 
 VariableValue = Union[str, int, float, dict, list, File]
 
-
 VARIABLE_PATTERN = re.compile(r"\{\{#([a-zA-Z0-9_]{1,50}(?:\.[a-zA-Z_][a-zA-Z0-9_]{0,29}){1,10})#\}\}")
 
 
@@ -131,7 +130,7 @@ class VariablePool(BaseModel):
             if attr not in {item.value for item in FileAttribute}:
                 return None
             value = self.get(selector)
-            if not isinstance(value, (FileSegment, NoneSegment)):
+            if not isinstance(value, FileSegment | NoneSegment):
                 return None
             if isinstance(value, FileSegment):
                 attr = FileAttribute(attr)

+ 2 - 2
api/core/workflow/nodes/loop/loop_node.py

@@ -1,6 +1,6 @@
 import logging
 from collections.abc import Generator, Mapping, Sequence
-from datetime import datetime, timezone
+from datetime import UTC, datetime
 from typing import Any, cast
 
 from configs import dify_config
@@ -80,7 +80,7 @@ class LoopNode(BaseNode[LoopNodeData]):
             thread_pool_id=self.thread_pool_id,
         )
 
-        start_at = datetime.now(timezone.utc).replace(tzinfo=None)
+        start_at = datetime.now(UTC).replace(tzinfo=None)
         condition_processor = ConditionProcessor()
 
         # Start Loop event

Plik diff jest za duży
+ 71 - 120
api/poetry.lock


+ 1 - 1
api/pyproject.toml

@@ -175,4 +175,4 @@ types-tqdm = "~4.67.0.20241221"
 optional = true
 [tool.poetry.group.lint.dependencies]
 dotenv-linter = "~0.5.0"
-ruff = "~0.9.9"
+ruff = "~0.11.0"

+ 4 - 4
api/services/dataset_service.py

@@ -949,7 +949,7 @@ class DocumentService:
                             ).first()
                             if document:
                                 document.dataset_process_rule_id = dataset_process_rule.id  # type: ignore
-                                document.updated_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+                                document.updated_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
                                 document.created_from = created_from
                                 document.doc_form = knowledge_config.doc_form
                                 document.doc_language = knowledge_config.doc_language
@@ -1916,7 +1916,7 @@ class SegmentService:
                 if cache_result is not None:
                     continue
                 segment.enabled = False
-                segment.disabled_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+                segment.disabled_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
                 segment.disabled_by = current_user.id
                 db.session.add(segment)
                 real_deal_segmment_ids.append(segment.id)
@@ -2008,7 +2008,7 @@ class SegmentService:
                         child_chunk.content = child_chunk_update_args.content
                         child_chunk.word_count = len(child_chunk.content)
                         child_chunk.updated_by = current_user.id
-                        child_chunk.updated_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+                        child_chunk.updated_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
                         child_chunk.type = "customized"
                         update_child_chunks.append(child_chunk)
             else:
@@ -2065,7 +2065,7 @@ class SegmentService:
             child_chunk.content = content
             child_chunk.word_count = len(content)
             child_chunk.updated_by = current_user.id
-            child_chunk.updated_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+            child_chunk.updated_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
             child_chunk.type = "customized"
             db.session.add(child_chunk)
             VectorService.update_child_chunk_vector([], [child_chunk], [], dataset)

+ 2 - 2
api/tasks/duplicate_document_indexing_task.py

@@ -51,7 +51,7 @@ def duplicate_document_indexing_task(dataset_id: str, document_ids: list):
             if document:
                 document.indexing_status = "error"
                 document.error = str(e)
-                document.stopped_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+                document.stopped_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
                 db.session.add(document)
         db.session.commit()
         return
@@ -80,7 +80,7 @@ def duplicate_document_indexing_task(dataset_id: str, document_ids: list):
                 db.session.commit()
 
             document.indexing_status = "parsing"
-            document.processing_started_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+            document.processing_started_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
             documents.append(document)
             db.session.add(document)
     db.session.commit()

+ 1 - 1
api/tasks/enable_segments_to_index_task.py

@@ -99,7 +99,7 @@ def enable_segments_to_index_task(segment_ids: list, dataset_id: str, document_i
             {
                 "error": str(e),
                 "status": "error",
-                "disabled_at": datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None),
+                "disabled_at": datetime.datetime.now(datetime.UTC).replace(tzinfo=None),
                 "enabled": False,
             }
         )

+ 3 - 3
api/tasks/retry_document_indexing_task.py

@@ -48,7 +48,7 @@ def retry_document_indexing_task(dataset_id: str, document_ids: list[str]):
             if document:
                 document.indexing_status = "error"
                 document.error = str(e)
-                document.stopped_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+                document.stopped_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
                 db.session.add(document)
                 db.session.commit()
             redis_client.delete(retry_indexing_cache_key)
@@ -76,7 +76,7 @@ def retry_document_indexing_task(dataset_id: str, document_ids: list[str]):
             db.session.commit()
 
             document.indexing_status = "parsing"
-            document.processing_started_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+            document.processing_started_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
             db.session.add(document)
             db.session.commit()
 
@@ -86,7 +86,7 @@ def retry_document_indexing_task(dataset_id: str, document_ids: list[str]):
         except Exception as ex:
             document.indexing_status = "error"
             document.error = str(ex)
-            document.stopped_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+            document.stopped_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
             db.session.add(document)
             db.session.commit()
             logging.info(click.style(str(ex), fg="yellow"))

+ 3 - 3
api/tasks/sync_website_document_indexing_task.py

@@ -46,7 +46,7 @@ def sync_website_document_indexing_task(dataset_id: str, document_id: str):
         if document:
             document.indexing_status = "error"
             document.error = str(e)
-            document.stopped_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+            document.stopped_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
             db.session.add(document)
             db.session.commit()
         redis_client.delete(sync_indexing_cache_key)
@@ -72,7 +72,7 @@ def sync_website_document_indexing_task(dataset_id: str, document_id: str):
         db.session.commit()
 
         document.indexing_status = "parsing"
-        document.processing_started_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+        document.processing_started_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
         db.session.add(document)
         db.session.commit()
 
@@ -82,7 +82,7 @@ def sync_website_document_indexing_task(dataset_id: str, document_id: str):
     except Exception as ex:
         document.indexing_status = "error"
         document.error = str(ex)
-        document.stopped_at = datetime.datetime.now(datetime.timezone.utc).replace(tzinfo=None)
+        document.stopped_at = datetime.datetime.now(datetime.UTC).replace(tzinfo=None)
         db.session.add(document)
         db.session.commit()
         logging.info(click.style(str(ex), fg="yellow"))

Niektóre pliki nie zostały wyświetlone z powodu dużej ilości zmienionych plików