Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
Open to Collab
14
2
19
AbstractPhila
PRO
AbstractPhil
Follow
andywu-kby's profile picture
Lionhearrtt's profile picture
liujiahiu's profile picture
78 followers
·
100 following
https://civitai.com/user/AbstractPhila
AbstractEyes
AI & ML interests
datasets, research papers, experimentation, vision, classification, text encoders, tokenization, llms, diffusion, distillation, and more.
Recent Activity
replied
to
their
post
about 9 hours ago
I'll attempt to expand the geolip-clip to full sequence context window to encompass sequential learning. https://huggingface.co/AbstractPhil/geolip-clip-vit-large-patch14-ctx576 The memory pod is specifically meant to tune everything based on final state pooling, which is fine if you aren't trying to actually use sequential utility. HOWEVER, there are many elemental biases that present themselves if attempting to USE the standard sequence of 77 in conjunction with this final pooled state. Even though the standard 77 is predominantly noise past token 10 it still houses considerable amounts of information in terms of utility, so this should be handled carefully. Zero-shot structures are a tricky structure to analyze, especially structures based on attention mechanisms instead of true sequential accumulation. I've noticed I need to watch them for quite a while before the real bugs show up. As it stands the token pool is essentially [B, 7+8, 768] for pools. This contains a robust and highly complex representation of useful accumulated bidirectional attention data, so it's quite powerful. I'll build a few prototypes and tap into some papers. I'll either come up with something or a reason why I didn't. The end result will either produce an anchor bank set of tokens [B, 15, 768] for pooling, or [B, 15, 77, 768] ideally - which should expand the sequence of the clip to 1,155 if successful. That doesn't necessarily mean this sequence will be more useful than the [b, 15, 768], but it will be representationally valid to the context window expansion. I wouldn't hold out for a single full-sequence option in a single day, that's a lot of moving parts to analyze, not to mention highly impractical to train with. A smaller dose of this information would be necessary for rapid prototyping so it'll likely be packaged as such. Well I spoke too soon. It's ready to play with. https://huggingface.co/AbstractPhil/geolip-clip-vit-large-patch14-ctx576-seq77
updated
a collection
about 10 hours ago
GeoLIP
published
a model
about 10 hours ago
AbstractPhil/geolip-clip-vit-bigG-patch14-ctx576-seq77
View all activity
Organizations
AbstractPhil
's models
134
Sort:Â Recently updated
AbstractPhil/omega-vit-l-reformed-fp32
0.4B
•
Updated
Apr 17, 2025
•
1
AbstractPhil/SD35-SIM-V1
Updated
Apr 16, 2025
•
4
AbstractPhil/t5xxl-unchained
Updated
Apr 7, 2025
•
4
•
4
AbstractPhil/SIM-OMEGA-PUBLIC-1
Updated
Apr 6, 2025
•
3
AbstractPhil/Beatrix
Updated
Apr 5, 2025
AbstractPhil/omega-vit-g-reformed
Updated
Apr 5, 2025
AbstractPhil/OMEGA-BIGASP
Updated
Apr 2, 2025
•
3
AbstractPhil/PONY-SIM-V4
Updated
Mar 28, 2025
•
1
AbstractPhil/SIM-V5
Updated
Mar 27, 2025
•
1
AbstractPhil/SDXL-SIM-REFINER
Updated
Mar 16, 2025
AbstractPhil/SDXL-SIM_NAI-VPRED
Updated
Mar 16, 2025
AbstractPhil/SDXL-Simulacrum-V3-1
0.2B
•
Updated
Mar 3, 2025
AbstractPhil/sdxl-interpolated
Text-to-Image
•
Updated
Feb 10, 2025
AbstractPhil/sdxl-interpolated-nai-xl-11
Text-to-Image
•
Updated
Feb 9, 2025
Previous
1
...
3
4
5
Next