Compare commits

..

11 Commits
merge ... main

Author SHA1 Message Date
Marsway 0928492ae4 fixing 2026-03-04 14:39:41 +08:00
Marsway 7d6ac687da fixing 2026-03-04 14:00:33 +08:00
Marsway 7262a33a93 fixing 2026-03-04 13:56:58 +08:00
Marsway 473343f548 fixing 2026-03-04 13:52:41 +08:00
Marsway ecc79128f6 update 2026-03-04 12:35:28 +08:00
Marsway 90a54a8dd8 fixing 2026-03-04 11:10:06 +08:00
Marsway a988696408 fixing 2026-03-04 11:02:45 +08:00
Marsway 7b0088c0bc Merge branch 'main' of https://git.marsway.red/liwei/Vastai-ConnectHub 2026-03-04 10:57:24 +08:00
Marsway 7f01d5b6b6 fixing 2026-03-04 10:57:03 +08:00
Li Wei 8d5f20a366 modify ignore 2026-03-04 02:37:31 +00:00
liwei 1f377a1db2 Merge pull request 'merge' (#1) from merge into main
Reviewed-on: #1
2026-03-04 02:08:52 +00:00
7 changed files with 461 additions and 6423 deletions

2
.env
View File

@ -3,7 +3,7 @@ DATA_DIR=/data
DB_URL=postgresql+psycopg://connecthub:connecthub_pwd_change_me@postgres:5432/connecthub
REDIS_URL=redis://redis:6379/0
FERNET_KEY_PATH=/data/fernet.key
DEV_MODE=1
DEV_MODE=0
LOG_DIR=/data/logs

1
.gitignore vendored
View File

@ -3,3 +3,4 @@
pgdata/
__pycache__/
*.pyc
logs/

View File

@ -53,6 +53,17 @@ class EhrClient(BaseClient):
self._token_type: str | None = None
self._token_expires_at: float | None = None
@staticmethod
def _normalize_token_type(token_type: str | None) -> str:
# 北森文档示例返回 token_type=bearer小写但鉴权头要求 "Bearer <token>"。
# 这里统一规范为首字母大写,避免服务端大小写敏感导致 401。
raw = str(token_type or "").strip()
if not raw:
return "Bearer"
if raw.lower() == "bearer":
return "Bearer"
return raw
def authenticate(self) -> str:
body: dict[str, Any] = {
"grant_type": self.grant_type,
@ -67,7 +78,7 @@ class EhrClient(BaseClient):
)
data = resp.json() if resp.content else {}
access_token = str(data.get("access_token", "") or "")
token_type = str(data.get("token_type", "") or "Bearer")
token_type = self._normalize_token_type(data.get("token_type"))
expires_in = int(data.get("expires_in", 0) or 0)
if not access_token:
raise RuntimeError("EHR authenticate failed (access_token missing)")
@ -88,7 +99,7 @@ class EhrClient(BaseClient):
def request(self, method: str, path: str, **kwargs: Any) -> httpx.Response: # type: ignore[override]
token = self._get_access_token()
token_type = self._token_type or "Bearer"
token_type = self._normalize_token_type(self._token_type)
headers = dict(kwargs.pop("headers", {}) or {})
headers["Authorization"] = f"{token_type} {token}"
@ -106,7 +117,7 @@ class EhrClient(BaseClient):
self._token_expires_at = None
token2 = self._get_access_token()
token_type2 = self._token_type or "Bearer"
token_type2 = self._normalize_token_type(self._token_type)
headers["Authorization"] = f"{token_type2} {token2}"
return super().request(method, path, headers=headers, **kwargs)

View File

@ -1,6 +1,6 @@
from __future__ import annotations
from datetime import datetime, timedelta
from datetime import datetime, timedelta, timezone
from uuid import uuid4
from fastapi import Request, Response
@ -8,14 +8,21 @@ from sqlalchemy import delete, select
from app.core.config import settings
from app.db.engine import get_session
from app.db.models import Session
from app.db.models import Session, User
SESSION_COOKIE_NAME = "session_id"
def _now_utc() -> datetime:
return datetime.utcnow()
return datetime.now(timezone.utc)
def _as_utc(dt: datetime) -> datetime:
# 兼容历史脏数据:若为 naive按 UTC 解释;若为 aware统一转换到 UTC。
if dt.tzinfo is None:
return dt.replace(tzinfo=timezone.utc)
return dt.astimezone(timezone.utc)
def create_session(user_id: int, request: Request) -> str:
@ -75,7 +82,7 @@ def get_current_user(request: Request) -> User | None:
if not record:
request.state.user = None
return None
if record.expires_at <= _now_utc():
if _as_utc(record.expires_at) <= _now_utc():
db.execute(delete(Session).where(Session.id == session_id))
db.commit()
request.state.user = None

File diff suppressed because it is too large Load Diff

View File

@ -1,7 +1,7 @@
from __future__ import annotations
import logging
from datetime import datetime
from datetime import datetime, timedelta
from typing import Any
from app.integrations.ehr import EhrClient
@ -41,13 +41,140 @@ class SyncEhrToOaApi:
self._client.close()
@staticmethod
def _to_api_datetime(value: datetime | str) -> str:
def _to_datetime(value: datetime | str | None) -> datetime:
if value is None:
return datetime.now()
if isinstance(value, datetime):
return value.strftime("%Y-%m-%dT%H:%M:%S")
return value
s = str(value).strip()
if not s:
raise ValueError("datetime string cannot be empty")
return s
if "T" in s:
return datetime.fromisoformat(s)
if " " in s:
return datetime.fromisoformat(s.replace(" ", "T"))
return datetime.strptime(s, "%Y-%m-%d")
@staticmethod
def _to_api_datetime(value: datetime | str | None) -> str:
return SyncEhrToOaApi._to_datetime(value).strftime("%Y-%m-%dT%H:%M:%S")
@staticmethod
def _iter_windows(start_dt: datetime, stop_dt: datetime, max_days: int = 90) -> list[tuple[datetime, datetime]]:
if stop_dt < start_dt:
raise ValueError("stop_time must be greater than or equal to start_time")
windows: list[tuple[datetime, datetime]] = []
cur = start_dt
max_delta = timedelta(days=max_days)
while cur < stop_dt:
nxt = min(cur + max_delta, stop_dt)
windows.append((cur, nxt))
cur = nxt
if not windows:
windows.append((start_dt, stop_dt))
return windows
def _get_all_by_time_window(
self,
*,
api_path: str,
api_name: str,
stop_time: datetime | str | None,
capacity: int,
time_window_query_type: int,
with_disabled: bool,
is_with_deleted: bool,
max_pages: int,
) -> dict[str, Any]:
if capacity <= 0 or capacity > 300:
raise ValueError("capacity must be in range [1, 300]")
if max_pages <= 0:
raise ValueError("max_pages must be > 0")
start_dt = datetime(2015, 1, 1, 0, 0, 0)
stop_dt = self._to_datetime(stop_time)
windows = self._iter_windows(start_dt=start_dt, stop_dt=stop_dt, max_days=90)
all_data: list[dict[str, Any]] = []
total_pages = 0
last_scroll_id = ""
for idx, (w_start, w_stop) in enumerate(windows, start=1):
start_time = self._to_api_datetime(w_start)
stop_time_s = self._to_api_datetime(w_stop)
scroll_id = ""
page = 0
window_total = 0
while True:
page += 1
total_pages += 1
if page > max_pages:
raise RuntimeError(f"scroll pages exceed max_pages={max_pages} in window index={idx}")
body: dict[str, Any] = {
"startTime": start_time,
"stopTime": stop_time_s,
"timeWindowQueryType": time_window_query_type,
"scrollId": scroll_id,
"capacity": capacity,
"withDisabled": with_disabled,
"isWithDeleted": is_with_deleted,
}
resp = self._client.request(
"POST",
api_path,
json=body,
headers={"Content-Type": "application/json"},
)
payload = resp.json() if resp.content else {}
code = str(payload.get("code", "") or "")
if code != "200":
message = payload.get("message")
raise RuntimeError(f"EHR {api_name} failed code={code!r} message={message!r}")
batch = payload.get("data") or []
if not isinstance(batch, list):
raise RuntimeError(f"EHR {api_name} invalid response: data is not a list")
all_data.extend([x for x in batch if isinstance(x, dict)])
total_val = payload.get("total")
if total_val is not None:
try:
window_total = int(total_val)
except (TypeError, ValueError):
pass
is_last_data = bool(payload.get("isLastData", False))
scroll_id = str(payload.get("scrollId", "") or "")
last_scroll_id = scroll_id
logger.info(
"EHR %s window=%s/%s page=%s batch=%s window_total=%s isLastData=%s",
api_name,
idx,
len(windows),
page,
len(batch),
window_total,
is_last_data,
)
if is_last_data:
break
return {
"startTime": self._to_api_datetime(start_dt),
"stopTime": self._to_api_datetime(stop_dt),
"total": len(all_data),
"pages": total_pages,
"count": len(all_data),
"data": all_data,
"lastScrollId": last_scroll_id,
"windowCount": len(windows),
}
def get_all_employees_with_record_by_time_window(
self,
@ -63,84 +190,18 @@ class SyncEhrToOaApi:
滚动查询员工 + 单条任职全量结果
固定起始时间
- 2001-01-01T00:00:00
- 2015-01-01T00:00:00
"""
if capacity <= 0 or capacity > 300:
raise ValueError("capacity must be in range [1, 300]")
if max_pages <= 0:
raise ValueError("max_pages must be > 0")
start_time = "2001-01-01T00:00:00"
stop_time_s = self._to_api_datetime(stop_time or datetime.now())
all_data: list[dict[str, Any]] = []
scroll_id = ""
total = 0
page = 0
while True:
page += 1
if page > max_pages:
raise RuntimeError(f"scroll pages exceed max_pages={max_pages}")
body: dict[str, Any] = {
"startTime": start_time,
"stopTime": stop_time_s,
"timeWindowQueryType": time_window_query_type,
"scrollId": scroll_id,
"capacity": capacity,
"withDisabled": with_disabled,
"isWithDeleted": is_with_deleted,
}
resp = self._client.request(
"POST",
"/TenantBaseExternal/api/v5/Employee/GetByTimeWindow",
json=body,
headers={"Content-Type": "application/json"},
return self._get_all_by_time_window(
api_path="/TenantBaseExternal/api/v5/Employee/GetByTimeWindow",
api_name="Employee.GetByTimeWindow",
stop_time=stop_time,
capacity=capacity,
time_window_query_type=time_window_query_type,
with_disabled=with_disabled,
is_with_deleted=is_with_deleted,
max_pages=max_pages,
)
payload = resp.json() if resp.content else {}
code = str(payload.get("code", "") or "")
if code != "200":
message = payload.get("message")
raise RuntimeError(f"EHR GetByTimeWindow failed code={code!r} message={message!r}")
batch = payload.get("data") or []
if not isinstance(batch, list):
raise RuntimeError("EHR GetByTimeWindow invalid response: data is not a list")
all_data.extend([x for x in batch if isinstance(x, dict)])
total_val = payload.get("total")
if total_val is not None:
try:
total = int(total_val)
except (TypeError, ValueError):
total = total
is_last_data = bool(payload.get("isLastData", False))
scroll_id = str(payload.get("scrollId", "") or "")
logger.info(
"EHR GetByTimeWindow page=%s batch=%s total=%s isLastData=%s",
page,
len(batch),
total,
is_last_data,
)
if is_last_data:
break
return {
"startTime": start_time,
"stopTime": stop_time_s,
"total": total,
"pages": page,
"count": len(all_data),
"data": all_data,
"lastScrollId": scroll_id,
}
def get_all_organizations_by_time_window(
self,
@ -156,84 +217,18 @@ class SyncEhrToOaApi:
滚动查询组织单元全量结果
固定起始时间
- 2001-01-01T00:00:00
- 2015-01-01T00:00:00
"""
if capacity <= 0 or capacity > 300:
raise ValueError("capacity must be in range [1, 300]")
if max_pages <= 0:
raise ValueError("max_pages must be > 0")
start_time = "2001-01-01T00:00:00"
stop_time_s = self._to_api_datetime(stop_time or datetime.now())
all_data: list[dict[str, Any]] = []
scroll_id = ""
total = 0
page = 0
while True:
page += 1
if page > max_pages:
raise RuntimeError(f"scroll pages exceed max_pages={max_pages}")
body: dict[str, Any] = {
"startTime": start_time,
"stopTime": stop_time_s,
"timeWindowQueryType": time_window_query_type,
"scrollId": scroll_id,
"capacity": capacity,
"withDisabled": with_disabled,
"isWithDeleted": is_with_deleted,
}
resp = self._client.request(
"POST",
"/TenantBaseExternal/api/v5/Organization/GetByTimeWindow",
json=body,
headers={"Content-Type": "application/json"},
return self._get_all_by_time_window(
api_path="/TenantBaseExternal/api/v5/Organization/GetByTimeWindow",
api_name="Organization.GetByTimeWindow",
stop_time=stop_time,
capacity=capacity,
time_window_query_type=time_window_query_type,
with_disabled=with_disabled,
is_with_deleted=is_with_deleted,
max_pages=max_pages,
)
payload = resp.json() if resp.content else {}
code = str(payload.get("code", "") or "")
if code != "200":
message = payload.get("message")
raise RuntimeError(f"EHR Organization.GetByTimeWindow failed code={code!r} message={message!r}")
batch = payload.get("data") or []
if not isinstance(batch, list):
raise RuntimeError("EHR Organization.GetByTimeWindow invalid response: data is not a list")
all_data.extend([x for x in batch if isinstance(x, dict)])
total_val = payload.get("total")
if total_val is not None:
try:
total = int(total_val)
except (TypeError, ValueError):
total = total
is_last_data = bool(payload.get("isLastData", False))
scroll_id = str(payload.get("scrollId", "") or "")
logger.info(
"EHR Organization.GetByTimeWindow page=%s batch=%s total=%s isLastData=%s",
page,
len(batch),
total,
is_last_data,
)
if is_last_data:
break
return {
"startTime": start_time,
"stopTime": stop_time_s,
"total": total,
"pages": page,
"count": len(all_data),
"data": all_data,
"lastScrollId": scroll_id,
}
def get_all_job_posts_by_time_window(
self,
@ -249,81 +244,15 @@ class SyncEhrToOaApi:
滚动查询职务全量结果
固定起始时间
- 2001-01-01T00:00:00
- 2015-01-01T00:00:00
"""
if capacity <= 0 or capacity > 300:
raise ValueError("capacity must be in range [1, 300]")
if max_pages <= 0:
raise ValueError("max_pages must be > 0")
start_time = "2001-01-01T00:00:00"
stop_time_s = self._to_api_datetime(stop_time or datetime.now())
all_data: list[dict[str, Any]] = []
scroll_id = ""
total = 0
page = 0
while True:
page += 1
if page > max_pages:
raise RuntimeError(f"scroll pages exceed max_pages={max_pages}")
body: dict[str, Any] = {
"startTime": start_time,
"stopTime": stop_time_s,
"timeWindowQueryType": time_window_query_type,
"scrollId": scroll_id,
"capacity": capacity,
"withDisabled": with_disabled,
"isWithDeleted": is_with_deleted,
}
resp = self._client.request(
"POST",
"/TenantBaseExternal/api/v5/JobPost/GetByTimeWindow",
json=body,
headers={"Content-Type": "application/json"},
return self._get_all_by_time_window(
api_path="/TenantBaseExternal/api/v5/JobPost/GetByTimeWindow",
api_name="JobPost.GetByTimeWindow",
stop_time=stop_time,
capacity=capacity,
time_window_query_type=time_window_query_type,
with_disabled=with_disabled,
is_with_deleted=is_with_deleted,
max_pages=max_pages,
)
payload = resp.json() if resp.content else {}
code = str(payload.get("code", "") or "")
if code != "200":
message = payload.get("message")
raise RuntimeError(f"EHR JobPost.GetByTimeWindow failed code={code!r} message={message!r}")
batch = payload.get("data") or []
if not isinstance(batch, list):
raise RuntimeError("EHR JobPost.GetByTimeWindow invalid response: data is not a list")
all_data.extend([x for x in batch if isinstance(x, dict)])
total_val = payload.get("total")
if total_val is not None:
try:
total = int(total_val)
except (TypeError, ValueError):
total = total
is_last_data = bool(payload.get("isLastData", False))
scroll_id = str(payload.get("scrollId", "") or "")
logger.info(
"EHR JobPost.GetByTimeWindow page=%s batch=%s total=%s isLastData=%s",
page,
len(batch),
total,
is_last_data,
)
if is_last_data:
break
return {
"startTime": start_time,
"stopTime": stop_time_s,
"total": total,
"pages": page,
"count": len(all_data),
"data": all_data,
"lastScrollId": scroll_id,
}

View File

@ -46,6 +46,97 @@ def _custom_prop_value(custom_props: Any, key: str | None) -> str:
return str(raw or "").strip()
def _to_bool_or_none(v: Any) -> bool | None:
if v is None:
return None
if isinstance(v, bool):
return v
s = str(v).strip().lower()
if s in ("1", "true", "yes", "y", "on"):
return True
if s in ("0", "false", "no", "n", "off", ""):
return False
return bool(v)
def _normalize_job_no(v: Any) -> str:
"""
工号标准化
- 去首尾空白去内部空格
- 数值型字符串如 123.0 -> 123常见于表单数字字段
- 统一大写便于大小写不敏感匹配
"""
s = str(v or "").strip()
if not s:
return ""
s = s.replace(" ", "")
try:
if "." in s and s.endswith(".0"):
i = int(float(s))
s = str(i)
except Exception:
pass
return s.upper()
def _extract_oa_row_id_and_fields(row: dict[str, Any]) -> tuple[int | None, dict[str, Any]]:
"""
兼容不同 OA export 返回结构提取
- row_id
- 字段字典key=fieldCode, value=单元格对象或值
"""
field_map: dict[str, Any] = {}
row_id: int | None = None
# 结构 AmasterData 直接是 {field0001: {value,showValue}, ...}
master = row.get("masterData")
if isinstance(master, dict):
for k, v in master.items():
if isinstance(k, str) and k.startswith("field"):
field_map[k] = v
for candidate in (row.get("id"), row.get("masterDataId"), master.get("id")):
if candidate is None:
continue
try:
row_id = int(str(candidate))
break
except Exception:
continue
# 结构 BmasterTable.record.fields = [{name,value,showValue}, ...]
master_table = row.get("masterTable")
if isinstance(master_table, dict):
record = master_table.get("record")
if isinstance(record, dict):
fields = record.get("fields")
if isinstance(fields, list):
for fld in fields:
if not isinstance(fld, dict):
continue
name = str(fld.get("name") or "").strip()
if name:
field_map[name] = fld
if row_id is None:
rid = record.get("id")
if rid is not None:
try:
row_id = int(str(rid))
except Exception:
pass
# 结构 C行级 fields 列表
row_fields = row.get("fields")
if isinstance(row_fields, list):
for fld in row_fields:
if not isinstance(fld, dict):
continue
name = str(fld.get("name") or "").strip()
if name:
field_map[name] = fld
return row_id, field_map
def _choose_better_record(current: dict[str, Any], candidate: dict[str, Any]) -> dict[str, Any]:
def _score(item: dict[str, Any]) -> str:
record = item.get("recordInfo") or {}
@ -114,6 +205,15 @@ class SyncEhrToOaFormJob(BaseJob):
rd_attr_custom_key = str(params.get("rd_attr_custom_key") or "").strip() or None
domain_custom_key = str(params.get("domain_account_custom_key") or "").strip() or None
verbose_trace = _to_bool_or_none(params.get("verbose_trace"))
if verbose_trace is None:
verbose_trace = True
preview_ehr_data = _to_bool_or_none(params.get("preview_ehr_data"))
if preview_ehr_data is None:
preview_ehr_data = True
preview_limit = int(params.get("preview_limit") or 20)
if preview_limit <= 0:
preview_limit = 20
seeyon = SeeyonClient(base_url=oa_base_url, rest_user=rest_user, rest_password=rest_password, loginName=login_name)
ehr = SyncEhrToOaApi(secret_params={"app_key": app_key, "app_secret": app_secret})
@ -139,6 +239,7 @@ class SyncEhrToOaFormJob(BaseJob):
# 3) 员工按工号归并(同工号保留“最新”记录)
ehr_by_job_no: dict[str, dict[str, Any]] = {}
ehr_by_job_no_norm: dict[str, dict[str, Any]] = {}
for item in emp_rows:
if not isinstance(item, dict):
continue
@ -150,15 +251,96 @@ class SyncEhrToOaFormJob(BaseJob):
continue
existing = ehr_by_job_no.get(job_no)
ehr_by_job_no[job_no] = item if existing is None else _choose_better_record(existing, item)
job_no_norm = _normalize_job_no(job_no)
if job_no_norm:
ex2 = ehr_by_job_no_norm.get(job_no_norm)
ehr_by_job_no_norm[job_no_norm] = item if ex2 is None else _choose_better_record(ex2, item)
logger.info(
"EHR 数据准备完成employee_rows=%s organization_rows=%s distinct_job_numbers=%s distinct_job_numbers_norm=%s",
len(emp_rows),
len(org_rows),
len(ehr_by_job_no),
len(ehr_by_job_no_norm),
)
if verbose_trace:
for job_no in list(ehr_by_job_no.keys()):
logger.info("EHR 工号明细raw=%s norm=%s", job_no, _normalize_job_no(job_no))
if preview_ehr_data:
logger.info("EHR 字段预览开始limit=%s", preview_limit)
count = 0
for job_no, item in ehr_by_job_no.items():
emp = item.get("employeeInfo") or {}
rec = item.get("recordInfo") or {}
if not isinstance(emp, dict):
emp = {}
if not isinstance(rec, dict):
rec = {}
org_oid = str(rec.get("oIdOrganization") or rec.get("oIdDepartment") or "").strip()
org = org_by_oid.get(org_oid, {})
company = str((org or {}).get("name") or "")
name = str(emp.get("name") or "")
rd_attr = _custom_prop_value(rec.get("customProperties"), rd_attr_custom_key) or _custom_prop_value(
emp.get("customProperties"), rd_attr_custom_key
)
place = str(rec.get("place") or "")
entry_date = _date_only(rec.get("entryDate"))
leave_date = _date_only(rec.get("lastWorkDate")) or "2099-12-31"
id_number = str(emp.get("iDNumber") or "")
hrbp = str((org or {}).get("hRBP") or "")
manager = str(rec.get("pOIdEmpAdmin") or "")
is_leaving = "" if _date_only(rec.get("lastWorkDate")) else ""
domain_account = _custom_prop_value(emp.get("customProperties"), domain_custom_key) or str(emp.get("_Name") or "")
logger.info(
"EHR 字段预览job_no=%s company=%s name=%s rd_attr=%s place=%s entry_date=%s leave_date=%s id_number=%s hrbp=%s manager=%s is_leaving=%s domain_account=%s",
job_no,
company,
name,
rd_attr,
place,
entry_date,
leave_date,
id_number,
hrbp,
manager,
is_leaving,
domain_account,
)
count += 1
if count >= preview_limit:
break
logger.info("EHR 字段预览结束printed=%s", count)
# 4) 导出 OA 表单,建立字段映射 + 工号到记录ID映射
exp_resp = seeyon.export_cap4_form_soap(templateCode=oa_template_code, senderLoginName=sender_login_name, rightId=oa_right_id)
exp_resp = seeyon.export_cap4_form_soap(
templateCode=oa_template_code,
senderLoginName=sender_login_name,
rightId=oa_right_id,
)
raw = exp_resp.text or ""
logger.info(
"OA export 返回status=%s content_length=%s template=%s",
exp_resp.status_code,
len(raw),
oa_template_code,
)
if raw:
logger.info("OA export 响应预览:%s", raw[:1000])
try:
payload = json.loads(raw) if raw else {}
except Exception as e: # noqa: BLE001
raise RuntimeError(f"OA export 响应不是有效 JSON: err={e!r} preview={raw[:500]!r}") from e
export_code = payload.get("code")
export_message = payload.get("message")
if export_code not in (None, 0, "0"):
raise RuntimeError(f"OA export failed code={export_code!r} message={export_message!r}")
outer = payload.get("data") or {}
form = outer.get("data") or {}
if not isinstance(form, dict):
raise RuntimeError("OA export invalid: data.data is not an object")
raise RuntimeError(
f"OA export invalid: data.data is not an object; payload_keys={list(payload.keys())[:20]}"
)
definition = form.get("definition") or {}
fields = definition.get("fields") or []
@ -209,40 +391,81 @@ class SyncEhrToOaFormJob(BaseJob):
oa_master_table_name = str(master_tbl.get("name") or "").strip()
if not oa_master_table_name:
raise RuntimeError("public_cfg.oa_master_table_name is required (cannot infer from OA export)")
logger.info(
"OA 表单解析完成template=%s master_table=%s form_rows=%s",
oa_template_code,
oa_master_table_name,
len(rows),
)
job_field_code = display_to_code["工号"]
oa_id_by_job_no: dict[str, int] = {}
oa_id_by_job_no_norm: dict[str, int] = {}
row_parse_miss = 0
for row in rows:
if not isinstance(row, dict):
continue
master = row.get("masterData") or {}
if not isinstance(master, dict):
continue
job_no = _cell_value(master.get(job_field_code))
row_id, field_map = _extract_oa_row_id_and_fields(row)
job_no = _cell_value(field_map.get(job_field_code))
if not job_no:
row_parse_miss += 1
if verbose_trace and row_parse_miss <= 20:
logger.info(
"OA 行解析未取到工号job_field=%s row_keys=%s field_keys_sample=%s",
job_field_code,
list(row.keys())[:20],
list(field_map.keys())[:20],
)
continue
row_id_raw = row.get("id")
if row_id_raw is None:
row_id_raw = row.get("masterDataId")
if row_id_raw is None:
row_id_raw = master.get("id")
if row_id_raw is None:
continue
try:
row_id = int(str(row_id_raw))
except Exception:
if row_id is None:
row_parse_miss += 1
if verbose_trace and row_parse_miss <= 20:
logger.info(
"OA 行解析未取到记录IDjob_no=%s row_keys=%s",
job_no,
list(row.keys())[:20],
)
continue
oa_id_by_job_no[job_no] = row_id
job_no_norm = _normalize_job_no(job_no)
if job_no_norm:
oa_id_by_job_no_norm[job_no_norm] = row_id
logger.info(
"OA 工号索引完成indexed_job_numbers=%s indexed_job_numbers_norm=%s parse_miss=%s",
len(oa_id_by_job_no),
len(oa_id_by_job_no_norm),
row_parse_miss,
)
if verbose_trace:
for job_no, row_id in list(oa_id_by_job_no.items()):
logger.info("OA 工号索引明细raw=%s norm=%s row_id=%s", job_no, _normalize_job_no(job_no), row_id)
# 5) 组装批量更新数据
data_list: list[dict[str, Any]] = []
not_found_in_oa = 0
unmatched_samples: list[str] = []
for job_no, item in ehr_by_job_no.items():
oa_record_id = oa_id_by_job_no.get(job_no)
matched_by = "raw"
if oa_record_id is None:
oa_record_id = oa_id_by_job_no_norm.get(_normalize_job_no(job_no))
matched_by = "normalized"
if oa_record_id is None:
not_found_in_oa += 1
if len(unmatched_samples) < 20:
unmatched_samples.append(job_no)
if verbose_trace:
logger.info("匹配失败job_no=%s norm=%s", job_no, _normalize_job_no(job_no))
continue
if verbose_trace:
logger.info(
"匹配成功job_no=%s norm=%s row_id=%s matched_by=%s",
job_no,
_normalize_job_no(job_no),
oa_record_id,
matched_by,
)
emp = item.get("employeeInfo") or {}
rec = item.get("recordInfo") or {}
@ -281,6 +504,10 @@ class SyncEhrToOaFormJob(BaseJob):
{"name": display_to_code["在离职"], "value": is_leaving, "showValue": is_leaving},
{"name": display_to_code["域账号"], "value": domain_account, "showValue": domain_account},
]
if verbose_trace:
logger.info("字段映射job_no=%s row_id=%s", job_no, oa_record_id)
for fld in fields_payload:
logger.info("字段映射明细job_no=%s field=%s value=%s", job_no, fld["name"], fld["value"])
data_list.append(
{
@ -295,20 +522,44 @@ class SyncEhrToOaFormJob(BaseJob):
"subTables": [],
}
)
logger.info(
"待更新数据准备完成prepared_updates=%s not_found_in_oa=%s",
len(data_list),
not_found_in_oa,
)
if not data_list:
raise RuntimeError(
"No updates prepared for OA batch-update (check jobNumber matching between EHR and OA, and form field mapping). "
f"unmatched_sample={unmatched_samples}"
)
# 6) 分批执行 batch-update
success_count = 0
failed_count = 0
failed_data: dict[str, str] = {}
do_trigger_bool = _to_bool_or_none(do_trigger)
for i in range(0, len(data_list), batch_size):
chunk = data_list[i : i + batch_size]
if verbose_trace:
logger.info("批量更新尝试chunk_index=%s chunk_size=%s", i // batch_size + 1, len(chunk))
for row in chunk:
try:
record = (((row or {}).get("masterTable") or {}).get("record") or {})
row_id = record.get("id")
fields = record.get("fields") or []
logger.info("批量更新行row_id=%s fields_count=%s", row_id, len(fields))
for fld in fields:
if isinstance(fld, dict):
logger.info("批量更新字段row_id=%s field=%s value=%s", row_id, fld.get("name"), fld.get("value"))
except Exception:
logger.info("批量更新行日志输出失败,已忽略")
resp = seeyon.batch_update_cap4_form_soap(
formCode=oa_form_code,
loginName=oa_login_name,
rightId=oa_right_id,
dataList=chunk,
uniqueFiled=[job_field_code],
doTrigger=bool(do_trigger) if do_trigger is not None else None,
doTrigger=do_trigger_bool,
)
rj = resp.json() if resp.content else {}
code = int(rj.get("code", -1))