|
|
|
|
|
import abc |
|
|
import argparse |
|
|
import csv |
|
|
import multiprocessing as mp |
|
|
import os |
|
|
import os.path as osp |
|
|
from pathlib import Path |
|
|
import copy as cp |
|
|
import random as rd |
|
|
import requests |
|
|
import shutil |
|
|
import subprocess |
|
|
import warnings |
|
|
import pandas as pd |
|
|
from collections import OrderedDict, defaultdict |
|
|
from multiprocessing import Pool, current_process |
|
|
from tqdm import tqdm |
|
|
import datetime |
|
|
import matplotlib.pyplot as plt |
|
|
from tabulate import tabulate |
|
|
from json import JSONDecoder |
|
|
from huggingface_hub import scan_cache_dir |
|
|
from huggingface_hub.utils._cache_manager import _scan_cached_repo |
|
|
from sty import fg, bg, ef, rs |
|
|
|
|
|
|
|
|
def modelscope_flag_set(): |
|
|
return os.environ.get('VLMEVALKIT_USE_MODELSCOPE', None) in ['1', 'True'] |
|
|
|
|
|
|
|
|
def process_punctuation(inText): |
|
|
import re |
|
|
outText = inText |
|
|
punct = [ |
|
|
';', r'/', '[', ']', '"', '{', '}', '(', ')', '=', '+', '\\', '_', '-', |
|
|
'>', '<', '@', '`', ',', '?', '!' |
|
|
] |
|
|
commaStrip = re.compile('(\d)(,)(\d)') |
|
|
periodStrip = re.compile('(?!<=\d)(\.)(?!\d)') |
|
|
for p in punct: |
|
|
if (p + ' ' in inText or ' ' + p in inText) or (re.search( |
|
|
commaStrip, inText) is not None): |
|
|
outText = outText.replace(p, '') |
|
|
else: |
|
|
outText = outText.replace(p, ' ') |
|
|
outText = periodStrip.sub('', outText, re.UNICODE) |
|
|
return outText |
|
|
|
|
|
def h2r(value): |
|
|
if value[0] == '#': |
|
|
value = value[1:] |
|
|
assert len(value) == 6 |
|
|
return tuple(int(value[i:i + 2], 16) for i in range(0, 6, 2)) |
|
|
|
|
|
def r2h(rgb): |
|
|
return '#%02x%02x%02x' % rgb |
|
|
|
|
|
def colored(s, color): |
|
|
if isinstance(color, str): |
|
|
if hasattr(fg, color): |
|
|
return getattr(fg, color) + s + fg.rs |
|
|
color = h2r(color) |
|
|
return fg(*color) + s + fg.rs |
|
|
|
|
|
def istype(s, type): |
|
|
if isinstance(s, type): |
|
|
return True |
|
|
try: |
|
|
return isinstance(eval(s), type) |
|
|
except Exception as _: |
|
|
return False |
|
|
|
|
|
def bincount(lst): |
|
|
bins = defaultdict(lambda: 0) |
|
|
for item in lst: |
|
|
bins[item] += 1 |
|
|
return bins |
|
|
|
|
|
def get_cache_path(repo_id, branch='main', repo_type='datasets'): |
|
|
try: |
|
|
if modelscope_flag_set(): |
|
|
from modelscope.hub.file_download import create_temporary_directory_and_cache |
|
|
if repo_type == 'datasets': |
|
|
repo_type = 'dataset' |
|
|
_, cache = create_temporary_directory_and_cache(model_id=repo_id, repo_type=repo_type) |
|
|
cache_path = cache.get_root_location() |
|
|
return cache_path |
|
|
else: |
|
|
from .file import HFCacheRoot |
|
|
cache_path = HFCacheRoot() |
|
|
org, repo_name = repo_id.split('/') |
|
|
repo_path = Path(osp.join(cache_path, f'{repo_type}--{org}--{repo_name}/')) |
|
|
hf_cache_info = _scan_cached_repo(repo_path=repo_path) |
|
|
revs = {r.refs: r for r in hf_cache_info.revisions} |
|
|
if branch is not None: |
|
|
revs = {refs: r for refs, r in revs.items() if branch in refs} |
|
|
rev2keep = max(revs.values(), key=lambda r: r.last_modified) |
|
|
return str(rev2keep.snapshot_path) |
|
|
except Exception as e: |
|
|
import logging |
|
|
logging.warning(f'{type(e)}: {e}') |
|
|
return None |
|
|
|
|
|
def proxy_set(s): |
|
|
import os |
|
|
for key in ['http_proxy', 'HTTP_PROXY', 'https_proxy', 'HTTPS_PROXY']: |
|
|
os.environ[key] = s |
|
|
|
|
|
def get_rank_and_world_size(): |
|
|
rank = int(os.environ.get('RANK', 0)) |
|
|
world_size = int(os.environ.get('WORLD_SIZE', 1)) |
|
|
return rank, world_size |
|
|
|
|
|
def splitlen(s, sym='/'): |
|
|
return len(s.split(sym)) |
|
|
|
|
|
def listinstr(lst, s): |
|
|
assert isinstance(lst, list) |
|
|
for item in lst: |
|
|
if item in s: |
|
|
return True |
|
|
return False |
|
|
|
|
|
def d2df(D): |
|
|
return pd.DataFrame({x: [D[x]] for x in D}) |
|
|
|
|
|
def cn_string(s): |
|
|
import re |
|
|
if re.search(u'[\u4e00-\u9fff]', s): |
|
|
return True |
|
|
return False |
|
|
|
|
|
try: |
|
|
import decord |
|
|
except ImportError: |
|
|
pass |
|
|
|
|
|
def timestr(granularity='second'): |
|
|
s = datetime.datetime.now().strftime('%Y%m%d%H%M%S') |
|
|
assert granularity in ['second', 'minute', 'hour', 'day'] |
|
|
if granularity == 'second': |
|
|
return s |
|
|
elif granularity == 'minute': |
|
|
return s[:-2] |
|
|
elif granularity == 'hour': |
|
|
return s[:-4] |
|
|
elif granularity == 'day': |
|
|
return s[:-6] |
|
|
|
|
|
def _minimal_ext_cmd(cmd, cwd=None): |
|
|
env = {} |
|
|
for k in ['SYSTEMROOT', 'PATH', 'HOME']: |
|
|
v = os.environ.get(k) |
|
|
if v is not None: |
|
|
env[k] = v |
|
|
env['LANGUAGE'] = 'C' |
|
|
env['LANG'] = 'C' |
|
|
env['LC_ALL'] = 'C' |
|
|
out = subprocess.Popen(cmd, stdout=subprocess.PIPE, env=env, cwd=cwd).communicate()[0] |
|
|
return out |
|
|
|
|
|
def githash(fallback='unknown', digits=8): |
|
|
if digits is not None and not isinstance(digits, int): |
|
|
raise TypeError('digits must be None or an integer') |
|
|
try: |
|
|
import vlmeval |
|
|
except ImportError as e: |
|
|
import logging |
|
|
logging.error(f'ImportError: {str(e)}') |
|
|
return fallback |
|
|
try: |
|
|
out = _minimal_ext_cmd(['git', 'rev-parse', 'HEAD'], cwd=vlmeval.__path__[0]) |
|
|
sha = out.strip().decode('ascii') |
|
|
if digits is not None: |
|
|
sha = sha[:digits] |
|
|
except OSError: |
|
|
sha = fallback |
|
|
return sha |
|
|
|
|
|
def dict_merge(dct, merge_dct): |
|
|
for k, _ in merge_dct.items(): |
|
|
if (k in dct and isinstance(dct[k], dict) and isinstance(merge_dct[k], dict)): |
|
|
dict_merge(dct[k], merge_dct[k]) |
|
|
else: |
|
|
dct[k] = merge_dct[k] |
|
|
|
|
|
def youtube_dl(idx): |
|
|
cmd = f'youtube-dl -f best -f mp4 "{idx}" -o {idx}.mp4' |
|
|
os.system(cmd) |
|
|
|
|
|
def run_command(cmd): |
|
|
if isinstance(cmd, str): |
|
|
cmd = cmd.split() |
|
|
return subprocess.check_output(cmd).decode() |
|
|
|
|
|
def load_env(): |
|
|
import logging |
|
|
logging.basicConfig( |
|
|
format='[%(asctime)s] %(levelname)s - %(filename)s: %(funcName)s - %(lineno)d: %(message)s', |
|
|
datefmt='%Y-%m-%d %H:%M:%S') |
|
|
|
|
|
try: |
|
|
import vlmeval |
|
|
except ImportError: |
|
|
logging.error('VLMEval is not installed. Failed to import environment variables from .env file. ') |
|
|
return |
|
|
pth = osp.realpath(vlmeval.__path__[0]) |
|
|
pth = osp.join(pth, '../.env') |
|
|
pth = osp.realpath(pth) |
|
|
if not osp.exists(pth): |
|
|
logging.error(f'Did not detect the .env file at {pth}, failed to load. ') |
|
|
return |
|
|
|
|
|
from dotenv import dotenv_values |
|
|
values = dotenv_values(pth) |
|
|
for k, v in values.items(): |
|
|
if v is not None and len(v): |
|
|
os.environ[k] = v |
|
|
logging.info(f'API Keys successfully loaded from {pth}') |
|
|
|
|
|
def pip_install_robust(package): |
|
|
import sys |
|
|
retry = 3 |
|
|
while retry > 0: |
|
|
try: |
|
|
package_base = package.split('=')[0] |
|
|
module = __import__(package) |
|
|
return True |
|
|
except ImportError: |
|
|
subprocess.check_call([sys.executable, '-m', 'pip', 'install', package]) |
|
|
retry -= 1 |
|
|
return False |
|
|
|
|
|
|
|
|
def version_cmp(v1, v2, op='eq'): |
|
|
from packaging import version |
|
|
import operator |
|
|
op_func = getattr(operator, op) |
|
|
return op_func(version.parse(v1), version.parse(v2)) |
|
|
|
|
|
|
|
|
def toliststr(s): |
|
|
if isinstance(s, str) and (s[0] == '[') and (s[-1] == ']'): |
|
|
return [str(x) for x in eval(s)] |
|
|
elif isinstance(s, str): |
|
|
return [s] |
|
|
elif isinstance(s, list): |
|
|
return [str(x) for x in s] |
|
|
raise NotImplementedError |
|
|
|
|
|
|
|
|
def extract_json_objects(text, decoder=JSONDecoder()): |
|
|
pos = 0 |
|
|
while True: |
|
|
match = text.find('{', pos) |
|
|
if match == -1: break |
|
|
try: |
|
|
result, index = decoder.raw_decode(text[match:]) |
|
|
yield result |
|
|
pos = match + index |
|
|
except ValueError: |
|
|
pos = match + 1 |
|
|
|
|
|
|
|
|
def get_gpu_memory(): |
|
|
import subprocess |
|
|
try: |
|
|
command = "nvidia-smi --query-gpu=memory.free --format=csv" |
|
|
memory_free_info = subprocess.check_output(command.split()).decode('ascii').split('\n')[:-1][1:] |
|
|
memory_free_values = [int(x.split()[0]) for i, x in enumerate(memory_free_info)] |
|
|
return memory_free_values |
|
|
except Exception as e: |
|
|
print(f'{type(e)}: {str(e)}') |
|
|
return [] |
|
|
|
|
|
|
|
|
def auto_split_flag(): |
|
|
flag = os.environ.get('AUTO_SPLIT', '0') |
|
|
if flag == '1': |
|
|
return True |
|
|
_, world_size = get_rank_and_world_size() |
|
|
try: |
|
|
import torch |
|
|
device_count = torch.cuda.device_count() |
|
|
if device_count > world_size and device_count % world_size == 0: |
|
|
return True |
|
|
else: |
|
|
return False |
|
|
except: |
|
|
return False |
|
|
|