Ksenia Se

Kseniase

AI & ML interests

None yet

Recent Activity

replied to their post 5 days ago
12 Types of JEPA Since Yann LeCun together with Randall Balestriero released a new paper on JEPA (Joint-Embedding Predictive Architecture), laying out its theory and introducing an efficient practical version called LeJEPA, we figured you might need even more JEPA. Here are 7 recent JEPA variants plus 5 iconic ones: 1. LeJEPA → https://huggingface.co/papers/2511.08544 Explains a full theory for JEPAs, defining the “ideal” JEPA embedding as an isotropic Gaussian, and proposes the SIGReg objective to push JEPA toward this ideal, resulting in practical LeJEPA 2. JEPA-T → https://huggingface.co/papers/2510.00974 A text-to-image model that tokenizes images and captions with a joint predictive Transformer, enhances fusion with cross-attention and text embeddings before training loss, and generates images by iteratively denoising visual tokens conditioned on text 3. Text-JEPA → https://huggingface.co/papers/2507.20491 Converts natural language into first-order logic, with a Z3 solver handling reasoning, enabling efficient, explainable QA with far lower compute than large LLMs 4. N-JEPA (Noise-based JEPA) → https://huggingface.co/papers/2507.15216 Connects self-supervised learning with diffusion-style noise by using noise-based masking and multi-level schedules, especially improving visual classification 5. SparseJEPA → https://huggingface.co/papers/2504.16140 Adds sparse representation learning to make embeddings more interpretable and efficient. It groups latent variables by shared semantic structure using a sparsity penalty while preserving accuracy 6. TS-JEPA (Time Series JEPA) → https://huggingface.co/papers/2509.25449 Adapts JEPA to time-series by learning latent self-supervised representations and predicting future latents for robustness to noise and confounders Read further below ↓ It you like it, also subscribe to the Turing Post: https://www.turingpost.com/subscribe
posted an update 5 days ago
12 Types of JEPA Since Yann LeCun together with Randall Balestriero released a new paper on JEPA (Joint-Embedding Predictive Architecture), laying out its theory and introducing an efficient practical version called LeJEPA, we figured you might need even more JEPA. Here are 7 recent JEPA variants plus 5 iconic ones: 1. LeJEPA → https://huggingface.co/papers/2511.08544 Explains a full theory for JEPAs, defining the “ideal” JEPA embedding as an isotropic Gaussian, and proposes the SIGReg objective to push JEPA toward this ideal, resulting in practical LeJEPA 2. JEPA-T → https://huggingface.co/papers/2510.00974 A text-to-image model that tokenizes images and captions with a joint predictive Transformer, enhances fusion with cross-attention and text embeddings before training loss, and generates images by iteratively denoising visual tokens conditioned on text 3. Text-JEPA → https://huggingface.co/papers/2507.20491 Converts natural language into first-order logic, with a Z3 solver handling reasoning, enabling efficient, explainable QA with far lower compute than large LLMs 4. N-JEPA (Noise-based JEPA) → https://huggingface.co/papers/2507.15216 Connects self-supervised learning with diffusion-style noise by using noise-based masking and multi-level schedules, especially improving visual classification 5. SparseJEPA → https://huggingface.co/papers/2504.16140 Adds sparse representation learning to make embeddings more interpretable and efficient. It groups latent variables by shared semantic structure using a sparsity penalty while preserving accuracy 6. TS-JEPA (Time Series JEPA) → https://huggingface.co/papers/2509.25449 Adapts JEPA to time-series by learning latent self-supervised representations and predicting future latents for robustness to noise and confounders Read further below ↓ It you like it, also subscribe to the Turing Post: https://www.turingpost.com/subscribe
replied to their post 12 days ago
7+ Main precision formats used in AI: Precision is very important in AI as it shapes how accurate and efficient models are. It controls how finely numbers are represented, approximating real-world values with formats like fixed-point and floating-point. A recent BF16 → FP16 study renewed attention to precision impact. Here are the main precision types used in AI, from full precision for training to ultra-low precision for inference: 1. FP32 (Float32): Standard full-precision float used in most training: 1 sign bit, 8 exponent bits, 23 mantissa bits. Default for backward-compatible training and baseline numerical stability 2. FP16 (Float16) → https://arxiv.org/abs/2305.10947v6 Half-precision float. It balances accuracy and efficiency. 1 sign bit, 5 exponent bits, 10 mantissa bits. Common on NVIDIA Tensor Cores and mixed-precision setups. There’s now a new wave of using it in reinforcement learning: https://www.turingpost.com/p/fp16 3. BF16 (BFloat16) → https://cloud.google.com/blog/products/ai-machine-learning/bfloat16-the-secret-to-high-performance-on-cloud-tpus Same dynamic range as FP32 but fewer mantissa bits: 1 sign bit, 8 exponent bits (same as FP32), 7 mantissa bits. It was developed by the research group Google Brain as part of their AI/ML infrastructure work at Google. Preferred on TPUs and modern GPUs 4. FP8 (E4M3 / E5M2) → https://proceedings.neurips.cc/paper_files/paper/2018/file/335d3d1cd7ef05ec77714a215134914c-Paper.pdf Emerging standard for training and inference on NVIDIA Hopper (H100) and Blackwell (B200) tensor cores and AMD MI300. Also supported in NVIDIA’s Transformer Engine: https://developer.nvidia.com/blog/floating-point-8-an-introduction-to-efficient-lower-precision-ai-training/ E4M3 = 4 exponent, 3 mantissa bits E5M2 = 5 exponent, 2 mantissa bits Read further below ⬇️ If you like this, also subscribe to the Turing post: https://www.turingpost.com/subscribe
View all activity

Organizations

Turing Post's profile picture Journalists on Hugging Face's profile picture Social Post Explorers's profile picture Hugging Face Discord Community's profile picture Sandbox's profile picture