LLFF / quantize.py
SlekLi's picture
Upload quantize.py
5b9b1ed verified
"""
3DGS Codebook Quantizer
========================
ไฝฟ็”จๅทฒ่ฎญ็ปƒๅฅฝ็š„ codebook ๅฏนๆ–ฐ็š„ 3DGS .ply ๆ–‡ไปถ่ฟ›่กŒ้‡ๅŒ–๏ผŒ
ๅฐ†่ฟž็ปญ็‰นๅพๆ˜ ๅฐ„ไธบ็ฆปๆ•ฃ็ดขๅผ•๏ผŒๅนถๅฏ้€‰ๅœฐ้‡ๅปบ้‡ๅŒ–ๅŽ็š„็‰นๅพๅ†™ๅ›ž .plyใ€‚
่พ“ๅ‡บ๏ผš
<scene>_quantized.npz โ€”โ€” ๅ››็ฑป็ดขๅผ• + ้‡ๅปบ่ฏฏๅทฎ็ปŸ่ฎก
<scene>_quantized.ply โ€”โ€” ๏ผˆๅฏ้€‰๏ผ‰็”จ codebook ้‡ๅปบ็‰นๅพๅŽๅ†™ๅ›ž็š„ๆ–ฐ .ply
"""
import os
import argparse
import numpy as np
from plyfile import PlyData, PlyElement
import time
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 1. PLY ่ฏปๅ–๏ผˆๅค็”จ่ฎญ็ปƒๆ—ถ็š„่ฏปๆณ•๏ผ‰
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def read_ply(ply_path: str) -> dict:
plydata = PlyData.read(ply_path)
vertex = plydata['vertex']
positions = np.stack([vertex['x'], vertex['y'], vertex['z']], axis=1)
opacities = vertex['opacity'][:, np.newaxis]
scales = np.stack([vertex['scale_0'], vertex['scale_1'],
vertex['scale_2']], axis=1)
rotations = np.stack([vertex['rot_0'], vertex['rot_1'],
vertex['rot_2'], vertex['rot_3']], axis=1)
dc = np.stack([vertex['f_dc_0'], vertex['f_dc_1'],
vertex['f_dc_2']], axis=1)
sh_keys = sorted(
[k for k in vertex.data.dtype.names if k.startswith('f_rest_')],
key=lambda s: int(s.split('_')[-1])
)
sh_rest = np.stack([vertex[k] for k in sh_keys], axis=1) \
if sh_keys else None
filter_3d = None
if 'filter_3D' in vertex.data.dtype.names:
filter_3d = vertex['filter_3D'][:, np.newaxis]
print(f"[read_ply] {os.path.basename(ply_path)}๏ผš{positions.shape[0]} ไธช้ซ˜ๆ–ฏ็‚น")
return {
'positions': positions,
'opacities': opacities,
'scales': scales,
'rotations': rotations,
'dc': dc,
'sh_rest': sh_rest,
'filter_3d': filter_3d,
'plydata': plydata,
'sh_keys': sh_keys,
}
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 2. ๅŠ ่ฝฝ codebook
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def load_codebook(codebook_dir: str, name: str):
"""
ๅŠ ่ฝฝๅ•ไธช codebook .npz๏ผŒ่ฟ”ๅ›ž codebook ็Ÿฉ้˜ต (K, D)ใ€‚
่ฎญ็ปƒๆ—ถไฟๅญ˜็š„ indices ๅฑžไบŽ่ฎญ็ปƒ้›†๏ผŒ้‡ๅŒ–ๆ—ถไธไฝฟ็”จใ€‚
"""
path = os.path.join(codebook_dir, f"{name}_codebook.npz")
if not os.path.exists(path):
raise FileNotFoundError(f"ๆ‰พไธๅˆฐ codebook ๆ–‡ไปถ๏ผš{path}")
npz = np.load(path)
codebook = npz['codebook'].astype(np.float32) # (K, D)
print(f"[load] {name}_codebook๏ผšK={codebook.shape[0]}, D={codebook.shape[1]}")
return codebook
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 3. ๆœ€่ฟ‘้‚ป้‡ๅŒ–๏ผˆๆ ธๅฟƒ๏ผ‰
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def quantize(features: np.ndarray, codebook: np.ndarray, name: str,
batch_size: int = 65536):
"""
ๅฏน features (N, D) ๅœจ codebook (K, D) ไธญๅšๆœ€่ฟ‘้‚ปๆœ็ดขใ€‚
้‡‡็”จๅˆ†ๆ‰น่ฎก็ฎ—้ฟๅ…ไธ€ๆฌกๆ€งๆž„้€  (N, K) ็š„ๅทจๅž‹็Ÿฉ้˜ตๆ’‘็ˆ†ๅ†…ๅญ˜ใ€‚
่ฟ”ๅ›ž๏ผš
indices : (N,) int32 ๆฏไธช็‚นๅฏนๅบ”็š„ codebook ็ดขๅผ•
reconstructed: (N, D) float32 ้‡ๅŒ–ๅŽ้‡ๅปบ็š„็‰นๅพ
"""
features = features.astype(np.float32)
N, D = features.shape
K = codebook.shape[0]
indices = np.empty(N, dtype=np.int32)
# โ”€โ”€ ๅˆ†ๆ‰นๆœ€่ฟ‘้‚ป โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# ๅˆฉ็”จๅฑ•ๅผ€็š„ L2 ่ท็ฆปๅ…ฌๅผ๏ผš
# ||x - c||^2 = ||x||^2 + ||c||^2 - 2 * x @ c^T
cb_norm2 = np.sum(codebook ** 2, axis=1) # (K,)
t0 = time.time()
for start in range(0, N, batch_size):
end = min(start + batch_size, N)
feat = features[start:end] # (B, D)
feat_norm2 = np.sum(feat ** 2, axis=1, keepdims=True) # (B, 1)
# (B, K) ่ท็ฆป็Ÿฉ้˜ต
dist2 = feat_norm2 + cb_norm2[np.newaxis, :] \
- 2.0 * (feat @ codebook.T)
indices[start:end] = np.argmin(dist2, axis=1)
elapsed = time.time() - t0
reconstructed = codebook[indices] # (N, D)
# โ”€โ”€ ่ฏฏๅทฎ็ปŸ่ฎก โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
diff = features - reconstructed
rmse = float(np.sqrt(np.mean(diff ** 2)))
max_e = float(np.abs(diff).max())
usage = len(np.unique(indices)) # ๅฎž้™…็”จๅˆฐๅคšๅฐ‘ไธช cluster
print(f"[{name:8s}] ้‡ๅŒ–ๅฎŒๆˆ {elapsed:.1f}s | "
f"RMSE={rmse:.6f} MaxErr={max_e:.6f} "
f"ไฝฟ็”จ {usage}/{K} ไธช cluster "
f"({100*usage/K:.1f}%)")
return indices, reconstructed, {'rmse': rmse, 'max_err': max_e, 'cluster_usage': usage}
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 4. ้‡ๅŒ–ๅ…จ้ƒจ็‰นๅพ
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def quantize_all(data: dict, codebook_dir: str):
"""
ๅŠ ่ฝฝๅ››ไธช codebook๏ผŒๅฏนๆ–ฐๅœบๆ™ฏ็š„้ซ˜ๆ–ฏ็‚น้€ไธ€้‡ๅŒ–ใ€‚
่ฟ”ๅ›ž๏ผš
results dict {name: {'indices', 'reconstructed', 'stats'}}
codebooks dict {name: np.ndarray}
"""
feature_map = {
'scale': data['scales'],
'rotation': data['rotations'],
'dc': data['dc'],
'sh': data['sh_rest'],
}
if data['sh_rest'] is None:
raise ValueError("PLY ไธญๆ—  f_rest_* ๅญ—ๆฎต๏ผŒๆ— ๆณ•้‡ๅŒ– SHใ€‚")
results = {}
codebooks = {}
for name, features in feature_map.items():
print(f"\n{'='*55}")
print(f" ้‡ๅŒ– [{name}] ็‰นๅพ็ปดๅบฆ: {features.shape[1]}")
print(f"{'='*55}")
cb = load_codebook(codebook_dir, name)
codebooks[name] = cb
idx, recon, stats = quantize(features, cb, name)
results[name] = {
'indices': idx,
'reconstructed': recon,
'stats': stats,
}
return results, codebooks
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 5. ไฟๅญ˜้‡ๅŒ–็ป“ๆžœ๏ผˆ็ดขๅผ• + ็ปŸ่ฎก๏ผ‰
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def save_quantized(save_path: str, data: dict, results: dict) -> None:
"""
ไฟๅญ˜้‡ๅŒ–ๅŽ็š„ๅ››็ฑป็ดขๅผ•ๅ’Œ็ปŸ่ฎกไฟกๆฏๅˆฐๅ•ไธช .npzใ€‚
ๆ–‡ไปถๅ†…ๅฎน๏ผš
scale_indices (N,) int32
rotation_indices (N,) int32
dc_indices (N,) int32
sh_indices (N,) int32
positions (N, 3) float32 ๅŽŸๅง‹ๅๆ ‡๏ผˆๆ–นไพฟๅŽ็ปญๅฏน้ฝ๏ผ‰
opacities (N, 1) float32
"""
save_dict = {
'positions': data['positions'].astype(np.float32),
'opacities': data['opacities'].astype(np.float32),
'scale_indices': results['scale']['indices'],
'rotation_indices': results['rotation']['indices'],
'dc_indices': results['dc']['indices'],
'sh_indices': results['sh']['indices'],
}
np.savez_compressed(save_path, **save_dict)
size_mb = os.path.getsize(save_path) / 1024 / 1024
print(f"\n[ไฟๅญ˜] ้‡ๅŒ–็ดขๅผ• โ†’ {save_path} ({size_mb:.2f} MB)")
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 6. ๏ผˆๅฏ้€‰๏ผ‰ๅ†™ๅ›ž้‡ๅŒ–้‡ๅปบ็š„ .ply
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def save_reconstructed_ply(
save_path: str,
data: dict,
results: dict,
) -> None:
"""
็”จ codebook ้‡ๅปบ็š„็‰นๅพๆ›ฟๆขๅŽŸๅง‹ๅ€ผ๏ผŒๅ†™ๅ‡บๆ–ฐ็š„ .ply ๆ–‡ไปถใ€‚
positions ๅ’Œ opacities ไฟๆŒไธๅ˜๏ผˆๆœช้‡ๅŒ–๏ผ‰ใ€‚
"""
plydata = data['plydata']
vertex = plydata['vertex']
sh_keys = data['sh_keys']
# โ”€โ”€ ๅ–ๅ‡บ้‡ๅปบๅ€ผ โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
scales_r = results['scale']['reconstructed'] # (N, 3)
rotations_r = results['rotation']['reconstructed'] # (N, 4)
dc_r = results['dc']['reconstructed'] # (N, 3)
sh_r = results['sh']['reconstructed'] # (N, 45)
# โ”€โ”€ ไฟฎๆ”น vertex ๆ•ฐ็ป„ โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# ๆณจๆ„๏ผšvertex.data ๆ˜ฏ็ป“ๆž„ๅŒ– numpy ๆ•ฐ็ป„๏ผŒ็›ดๆŽฅๆŒ‰ๅญ—ๆฎต่ต‹ๅ€ผ
arr = vertex.data.copy()
arr['scale_0'] = scales_r[:, 0]
arr['scale_1'] = scales_r[:, 1]
arr['scale_2'] = scales_r[:, 2]
arr['rot_0'] = rotations_r[:, 0]
arr['rot_1'] = rotations_r[:, 1]
arr['rot_2'] = rotations_r[:, 2]
arr['rot_3'] = rotations_r[:, 3]
arr['f_dc_0'] = dc_r[:, 0]
arr['f_dc_1'] = dc_r[:, 1]
arr['f_dc_2'] = dc_r[:, 2]
for i, key in enumerate(sh_keys):
arr[key] = sh_r[:, i]
# โ”€โ”€ ๅ†™ๅ‡บๆ–ฐ ply โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
new_vertex = PlyElement.describe(arr, 'vertex')
new_plydata = PlyData([new_vertex], text=plydata.text)
new_plydata.write(save_path)
size_mb = os.path.getsize(save_path) / 1024 / 1024
print(f"[ไฟๅญ˜] ้‡ๅปบ .ply โ†’ {save_path} ({size_mb:.2f} MB)")
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 7. ๆ‰“ๅฐๆฑ‡ๆ€ป็ปŸ่ฎก
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def print_summary(results: dict) -> None:
print(f"\n{'='*55}")
print(f" ้‡ๅŒ–ๆฑ‡ๆ€ป")
print(f"{'='*55}")
print(f" {'็‰นๅพ':<10} {'RMSE':>10} {'MaxErr':>10} {'Clusterไฝฟ็”จ็އ':>14}")
print(f" {'-'*46}")
for name, res in results.items():
s = res['stats']
print(f" {name:<10} {s['rmse']:>10.6f} {s['max_err']:>10.6f} "
f" {s['cluster_usage']:>5} / {len(np.unique(res['indices'])):>5}"
f" ({100*s['cluster_usage']/s['cluster_usage']:.0f}%)")
print(f"{'='*55}")
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
# 8. CLI ๅ…ฅๅฃ
# โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
def parse_args():
parser = argparse.ArgumentParser(
description="็”จๅทฒ่ฎญ็ปƒ็š„ codebook ้‡ๅŒ–ๆ–ฐ็š„ 3DGS .ply ๆ–‡ไปถ"
)
parser.add_argument('ply_path', type=str,
help='ๅพ…้‡ๅŒ–็š„ 3DGS .ply ๆ–‡ไปถ่ทฏๅพ„')
parser.add_argument('--codebook_dir', type=str, default='./codebooks',
help='ๅญ˜ๆ”พๅ››ไธช *_codebook.npz ็š„็›ฎๅฝ•๏ผˆ้ป˜่ฎค๏ผš./codebooks๏ผ‰')
parser.add_argument('--save_dir', type=str, default='./quantized',
help='้‡ๅŒ–็ป“ๆžœ่พ“ๅ‡บ็›ฎๅฝ•๏ผˆ้ป˜่ฎค๏ผš./quantized๏ผ‰')
parser.add_argument('--save_ply', action='store_true',
help='ๅŒๆ—ถ่พ“ๅ‡บ็”จ codebook ้‡ๅปบ็‰นๅพๅŽ็š„ .ply ๆ–‡ไปถ')
return parser.parse_args()
if __name__ == '__main__':
args = parse_args()
os.makedirs(args.save_dir, exist_ok=True)
# โ”€โ”€ ่ฏปๅ–ๆ–ฐๅœบๆ™ฏ โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
data = read_ply(args.ply_path)
# โ”€โ”€ ้‡ๅŒ– โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
results, codebooks = quantize_all(data, args.codebook_dir)
# โ”€โ”€ ๆ‰“ๅฐๆฑ‡ๆ€ป โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
print_summary(results)
# โ”€โ”€ ไฟๅญ˜็ดขๅผ• โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
scene_name = os.path.splitext(os.path.basename(args.ply_path))[0]
npz_path = os.path.join(args.save_dir, f"{scene_name}_quantized.npz")
save_quantized(npz_path, data, results)
# โ”€โ”€ ๏ผˆๅฏ้€‰๏ผ‰ๅ†™ๅ›ž้‡ๅปบ ply โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€
if args.save_ply:
ply_out = os.path.join(args.save_dir, f"{scene_name}_reconstructed.ply")
save_reconstructed_ply(ply_out, data, results)
print("\nๅ…จ้ƒจๅฎŒๆˆ๏ผ")