Cool! any chance at sharing a repo so others can play around with it? I'd love to give it a try! ๐
Tyler Williams PRO
unmodeled-tyler
AI & ML interests
AI research engineer & solo operator of VANTA Research/Quanta Intellect
Recent Activity
repliedto omarkamali's post about 9 hours ago
I just might have cracked tokenizer-free LLMs. No vocab, no softmax.
I'm training a 22M params LLM rn to test this "thing" and it's able to formulate coherent sentences ๐คฏ
Bear in mind, this is a completely new, tokenizer-free LLM architecture with built-in language universality.
Check the explainer video to understand what's happening. Feedback welcome on this approach!
reacted to karstenskyt's post with ๐ฅ about 9 hours ago
๐ ๐๐ฎ๐๐ป๐ฐ๐ต๐ถ๐ป๐ด ๐๐ต๐ฒ ๐๐/๐ ๐ ๐ช๐ผ๐ฟ๐ธ๐ณ๐น๐ผ๐๐ ๐๐ฎ๐๐ต๐ฏ๐ผ๐ฎ๐ฟ๐ฑ
Now that our Taipy architecture is humming along on Hugging Face Spaces, we just shipped the most complex feature of the (๐๐ช๐จ๐ฉ๐ต! ๐๐ถ๐น๐ถ๐ณ๐บ!) ๐๐ข๐ฌ๐ฆ๐ฉ๐ฐ๐ถ๐ด๐ฆ to date: the ๐๐/๐ ๐ ๐ช๐ผ๐ฟ๐ธ๐ณ๐น๐ผ๐๐ ๐๐ฎ๐๐ต๐ฏ๐ผ๐ฎ๐ฟ๐ฑ.
Managing 16 different machine learning pipelines (from Expected Goals to Space Creation) across Databricks Serverless and HF Jobs is a logistical challenge. To solve this, we built a dynamic operations center (the 13th page in our app).
It features:
ย ย โข ๐๐ป ๐ถ๐ป๐๐ฒ๐ฟ๐ฎ๐ฐ๐๐ถ๐๐ฒ ๐ฑ๐ฒ๐ฝ๐ฒ๐ป๐ฑ๐ฒ๐ป๐ฐ๐ ๐๐๐: Powered by Cytoscape.js, it visually maps exactly how our models and data grids feed into each other.
ย ย โข ๐ฅ๐ฒ๐ฎ๐น-๐๐ถ๐บ๐ฒ ๐บ๐ผ๐ป๐ถ๐๐ผ๐ฟ๐ถ๐ป๐ด: Tracks run volumes and data freshness SLAs across the entire platform.
ย ย โข ๐ ๐ฏ-๐๐ถ๐ฒ๐ฟ ๐ต๐๐ฏ๐ฟ๐ถ๐ฑ ๐ฐ๐ผ๐๐ ๐ฒ๐ป๐ด๐ถ๐ป๐ฒ: Merges "cold" Databricks billing data with "warm/hot" live HF Jobs estimates to give a unified view of pipeline expenses.
Check out the live interactive graph here:
https://huggingface.co/spaces/luxury-lakehouse/soccer-analytics-app posted an update about 20 hours ago
PSA: LiteLLM has been compromised on PyPI - if you have it installed, CHECK NOW.
LiteLLM is used as a dependency in A LOT of AI tooling, so there's a pretty good chance that you have it installed somewhere on your machine (my instance was part of Hermes Agent, but I was unaffected by the hack)
Versions 1.82.7 & 1.82.8 on PyPI have been compromised with a multi-stage credential stealer.
- Version 1.82.8 uses a .pth file that executes on EVERY python process startup. You don't even need to import litellm. Just having it installed is enough.
- The payload harvests SSH keys, .env files, AWS/GCP/Azure credentials, Kubernetes configs, database passwords, crytpo wallets, shell history - basically every secret on your machine.
- Stolen data is encrypted with a hardcoded RSA key and exfiltrated to a domain that is NOT part of a legitimate litellm infrastructure.
- If you're running Kubernetes, it attempts lateral movement across the entire cluster.
- The C2 is hosted on the Internet Computer blockchain, making it essentially impossible to take down.
This is part of a coordinated campaign by a threat actor called TeamPCP who have also hit Trivy (Aqua Security), Checkmarx KICS, and multiple npm packages in the last week ALONE.
What to do:
1. Run 'pip show litellm' in every environment you have
2. If you're on 1.82.7 or 1.82.8 - rotate EVERY secret on that machine immediately.
3. Check for persistence artifacts ~/.config/sysmon/sysmon.py & ~/.config/systemd/user/sysmon.service
I was lucky in this case that my litellm version was out of date, but if you've installed litellm as a dependency in ANY package within the last 24ish hours, you're gonna want to check.
SOURCES
https://futuresearch.ai/blog/litellm-pypi-supply-chain-attack/
Same group, different attack a couple of days ago: https://www.stepsecurity.io/blog/canisterworm-how-a-self-propagating-npm-worm-is-spreading-backdoors-across-the-ecosystem