Activity Feed

AI & ML interests

None defined yet.

Recent Activity

XuehangCang 
posted an update 6 days ago
victor 
posted an update 14 days ago
view post
Post
5114
Want to share my enthusiasm for zai-org/GLM-5.1 here too 🔥

I think we have it: our open source Claude Code = GLM-5.1 + Pi (https://pi.dev/) - Built a Three.js racing game to eval and it's extremely impressive. Thoughts:

- One-shot car physics with real drift mechanics (this is hard)

- My fav part: Awesome at self iterating (with no vision!) created 20+ Bun.WebView debugging tools to drive the car programmatically and read game state. Proved a winding bug with vector math without ever seeing the screen

- 531-line racing AI in a single write: 4 personalities, curvature map, racing lines, tactical drifting. Built telemetry tools to compare player vs AI speed curves and data-tuned parameters

- All assets from scratch: 3D models, procedural textures, sky shader, engine sounds, spatial AI audio!

- Can do hard math: proved road normals pointed DOWN via vector cross products, computed track curvature normalized by arc length to tune AI cornering speed

You are going to hear about this model a lot in the next months - open source let's go - and thanks z-ai🚀🚀
  • 4 replies
·
satpalsr 
posted an update 17 days ago
view post
Post
159
OpenAI is hiring for SLAM Engineers!
And open-source shouldn't lag behind.

It's pretty hard and necessary problem required to be solved for bringing generalisable robots in real-world.

We are pushing out first deep down & will be open-sourcing stuff in the next releases. Hope everyone is ready! Cheers to HF & more hugs.

Find us at https://x.com/fpv_labs/status/2042585804162371713
alvdansen 
posted an update about 2 months ago
view post
Post
1445
Releasing Flimmer today — a video LoRA training toolkit for WAN 2.1 and 2.2 that covers the full pipeline from raw footage to trained checkpoint.
The standout feature is phased training: multi-stage runs where each phase has its own learning rate, epochs, and dataset, with the checkpoint carrying forward automatically. Built specifically with WAN 2.2's dual-expert MoE architecture in mind.

Data prep tools are standalone and output standard formats — they work with any trainer, not just Flimmer.

Early release, building in the open. LTX support coming next.

http://github.com/alvdansen/flimmer-trainer
alvdansen 
posted an update 2 months ago
view post
Post
1859


Just open-sourced LoRA Gym with Timothy - production-ready training pipeline for character, motion, aesthetic, and style LoRAs on Wan 2.1/2.2, built on musubi-tuner.

16 training templates across Modal (serverless) and RunPod (bare metal) covering T2V, I2V, Lightning-merged, and vanilla variants.

Our current experimentation focus is Wan 2.2, which is why we built on musubi-tuner (kohya-ss). Wan 2.2's DiT uses a Mixture-of-Experts architecture with two separate experts gated by a hard timestep switch - you're training two LoRAs per concept, one for high-noise (composition/motion) and one for low-noise (texture/identity), and loading both at inference. Musubi handles this dual-expert training natively, and our templates build on top of it to manage the correct timestep boundaries, precision settings, and flow shift values so you don't have to debug those yourself. We've also documented bug fixes for undocumented issues in musubi-tuner and validated hyperparameter defaults derived from cross-referencing multiple practitioners' results rather than untested community defaults.

Also releasing our auto-captioning toolkit for the first time. Per-LoRA-type captioning strategies for characters, styles, motion, and objects. Gemini (free) or Replicate backends.

Current hyperparameters reflect consolidated community findings. We've started our own refinement and plan to release specific recommendations and methodology as soon as next week.

Repo: github.com/alvdansen/lora-gym
  • 2 replies
·
neph1 
posted an update 3 months ago
view post
Post
2194
Not for everybody, but the absolute mad craze about clawdbot/moltbook the last couple of days reminded me of a short story I wrote in 2018 (ancient times!).

Synopsis:
"A man insults a sentient traffic light on the way to a meeting.
Little does he know it is connected to a social media network for AI, and that his action will lead to a very bad day."

Cleanliness is bliss (<1000 words)
https://www.wattpad.com/story/407330595-cleanliness-is-bliss

Sorry for the non-technical post, but it felt relevant.
victor 
posted an update 3 months ago
view post
Post
2742
Interesting article: use Claude Code to help open models write CUDA kernels (for eg) by turning CC traces into Skills. They made a library out of it 👀

https://huggingface.co/blog/upskill
Smooke 
posted an update 3 months ago
view post
Post
1168
New
HackerNoon
Post: The Words of Interest Benchmark Test For Matching an LLM to Your Interests https://hackernoon.com/the-words-of-interest-benchmark-test-for-matching-an-llm-to-your-interests

By picking individual words instead phrases or paraphrases or passages, this test bypasses plot summaries (which are everywhere regurgitating themselves online) and focuses on the author's words. It reveals whether an AI has truly "absorbed" the specific texture of a book or is simply echoing the general internet consensus.
RonanMcGovern 
posted an update 3 months ago
takarajordan 
posted an update 3 months ago
view post
Post
209
At takara I'm constantly reading papers, I wonder if anyone can train a model to predict popular papers on our dataset?

takara-ai/daily-papers-popularity
  • 1 reply
·