Ugiat Technologies

238 posts

Ugiat Technologies banner
Ugiat Technologies

Ugiat Technologies

@UgiatTech

Specialist in Computer Vision, Image Processing, Machine Learning (Deep Learning), Audio Signal Processing. Releasing https://t.co/vIQ9OrhRif low-cost dubbing service

Barcelona, España Katılım Temmuz 2015
358 Takip Edilen136 Takipçiler
Ugiat Technologies retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
New Andrej Karpathy interview: "To get the most out of the tools that have become available now, you have to remove yourself as the bottleneck. You cannot be there to prompt the next thing. You need to take yourself outside the loop. You have to arrange things such that they are completely autonomous. The more you can maximize your token throughput and not be in the loop, the better. This is the goal. So, I kind of mentioned that the name of the game now is to increase your leverage. I put in very few tokens just once in a while, and a huge amount of stuff happens on my behalf." --- From @NoPriorsPod YT channel (link in comment)
English
56
151
1.5K
147.5K
Ugiat Technologies
Ugiat Technologies@UgiatTech·
Spending hours watching #longvideos just to find a few key insights? There’s a smarter way. With DoblAI, instantly generate accurate #transcripts, #translations and #subtitles — fully synced and ready to use. Work faster, repurpose content easily, and make every video count
English
0
0
0
36
Ugiat Technologies retweetledi
PyTorch
PyTorch@PyTorch·
Excited to share our latest deep-dive on @NVIDIA's DGX Spark AI-PC! Discover how we unlocked advanced reasoning in Llama 8B through full fine-tuning—entirely offline, thanks to DGX Spark’s unified 128GB memory. Learn how synthetic data and chain-of-thought prompts are pushing the boundaries of local LLMs. 🖇️ Read the full recipe and results on the PyTorch Foundation blog: hubs.la/Q041mvSQ0 👥 @bhutanisanyam1 (PyTorch Meta), Hamid Shojanazeri (PyTorch Meta), Clement Anthonioz Blanc (Meta) #PyTorch #DGXSpark #AI #LLM #FineTuning
PyTorch tweet media
English
7
4
72
7.6K
Ugiat Technologies retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
A solid 65-page long paper from Stanford, Princeton, Harvard, University of Washington, and many other top univ. Says that almost all advanced AI agent systems can be understood as using just 4 basic ways to adapt, either by updating the agent itself or by updating its tools. It also positions itself as the first full taxonomy for agentic AI adaptation. Agentic AI means a large model that can call tools, use memory, and act over multiple steps. Adaptation here means changing either the agent or its tools using a kind of feedback signal. In A1, the agent is updated from tool results, like whether code ran correctly or a query found the answer. In A2, the agent is updated from evaluations of its outputs, for example human ratings or automatic checks of answers and plans. In T1, retrievers that fetch documents or domain models for specific fields are trained separately while a frozen agent just orchestrates them. In T2, the agent stays fixed but its tools are tuned from agent signals, like which search results or memory updates improve success. The survey maps many recent systems into these 4 patterns and explains trade offs between training cost, flexibility, generalization, and modular upgrades.
Rohan Paul tweet media
English
29
225
1.2K
71K
Ugiat Technologies retweetledi
Akshay 🚀
Akshay 🚀@akshay_pachaar·
Google just dropped "Attention is all you need (V2)" This paper could solve AI's biggest problem: Catastrophic forgetting. When AI models learn something new, they tend to forget what they previously learned. Humans don't work this way, and now Google Research has a solution. Nested Learning. This is a new machine learning paradigm that treats models as a system of interconnected optimization problems running at different speeds - just like how our brain processes information. Here's why this matters: LLMs don't learn from experiences; they remain limited to what they learned during training. They can't learn or improve over time without losing previous knowledge. Nested Learning changes this by viewing the model's architecture and training algorithm as the same thing - just different "levels" of optimization. The paper introduces Hope, a proof-of-concept architecture that demonstrates this approach: ↳ Hope outperforms modern recurrent models on language modeling tasks ↳ It handles long-context memory better than state-of-the-art models ↳ It achieves this through "continuum memory systems" that update at different frequencies This is similar to how our brain manages short-term and long-term memory simultaneously. We might finally be closing the gap between AI and the human brain's ability to continually learn. I've shared link to the paper in the next tweet!
Akshay 🚀 tweet media
English
257
1K
6K
511.4K
Ugiat Technologies retweetledi
PyTorch
PyTorch@PyTorch·
Need to optimize applications with large models and stretched memory resources? Learn how to accelerate large-scale LLM inference and Cache Offload with CPU-GPU Memory Sharing in NVIDIA’s recent developer tech blog: 📎: hubs.la/Q03V7v_q0 #PyTorch #OpenSourceAI #AI #Inference #Innovation
English
3
12
97
11.9K
Ugiat Technologies retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
NVIDIA's brilliant paper gives a lot of details and actionable techniques. 🎯 Small Language Models (SLMs) not LLMs are the real future of agentic AI. They gave a recipe for swapping out the large models with SLMs without breaking anything and show that 40%‑70% of calls in open agents could switch today. It argues that SLMs, already match big models on many routine agent tasks, cost far less to run, and slot neatly into mixed‑model pipelines, so most agent calls should shift to SLMs. 🗝️ The central claim SLMs that fit on a laptop already handle the narrow language chores inside most agents. Because those chores rarely need open‑ended chat, the authors say an SLM‑first design is the natural default, and large models become occasional helpers. @NVIDIAAIDev 👏
Rohan Paul tweet media
English
22
212
1.1K
78.7K
Ugiat Technologies retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
This github repo is a goldmine. 3.4K Starts ⭐️ in 4 days. end-to-end, code-first tutorials covering every layer of production-grade GenAI agents, guiding you from spark to scale with proven patterns and reusable blueprints for real-world launches.
Rohan Paul tweet media
English
26
368
3.1K
352.5K
Ugiat Technologies retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
Anthropic free masterclass on prompt engineering is one of the best 👌 Teaches you how to write effective prompts for Claude models. You’ll start with basic prompt structure, clarity, and role assignment in three beginner chapters. Intermediate lessons cover separating data from instructions, formatting outputs, and step-by-step reasoning (“precognition”), with hands-on exercises to reinforce each topic. Advanced chapters tackle hallucination avoidance and building complex prompts for chatbots, legal, financial, and coding use cases. Interactive “Example Playground” sections let you experiment in real time and refer to an answer key for guidance.
Rohan Paul tweet media
English
9
111
954
80.6K
Ugiat Technologies retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
Making a model smarter by forcing it to argue with itself actually works.
Rohan Paul tweet media
English
101
212
4.1K
438.3K
Ugiat Technologies retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
Some nice collection of resources for LLMs
Rohan Paul tweet media
English
14
139
1.1K
103.7K
Ugiat Technologies retweetledi
Vive con Propósito.
Vive con Propósito.@PropositoyVida·
¡Llegaste hasta el final! Si te ha gustado este hilo, por favor retuitea y dale me gusta al primer tweet y sígueme @PropositoyVida Te recomiendo guardar el hilo, para que puedas consultarlo cuando quieras. Ayúdame a que más personas lo lean. ¡Muchas gracias por tu apoyo!
Español
64
806
5.1K
427.2K
Ugiat Technologies retweetledi
Ruben Tous
Ruben Tous@rubentous1·
Seeing the level of understanding of the physical world of the latest vision models, the question is not if this is intelligence but if our intelligence is actually this. x.com/_akhaliq/statu…
AK@_akhaliq

PhysGen Rigid-Body Physics-Grounded Image-to-Video Generation We present PhysGen, a novel image-to-video generation method that converts a single image and an input condition (e.g., force and torque applied to an object in the image) to produce a realistic, physically plausible, and temporally consistent video. Our key insight is to integrate model-based physical simulation with a data-driven video generation process, enabling plausible image-space dynamics. At the heart of our system are three core components: (i) an image understanding module that effectively captures the geometry, materials, and physical parameters of the image; (ii) an image-space dynamics simulation model that utilizes rigid-body physics and inferred parameters to simulate realistic behaviors; and (iii) an image-based rendering and refinement module that leverages generative video diffusion to produce realistic video footage featuring the simulated motion. The resulting videos are realistic in both physics and appearance and are even precisely controllable, showcasing superior results over existing data-driven image-to-video generation works through quantitative comparison and comprehensive user study. PhysGen's resulting videos can be used for various downstream applications, such as turning an image into a realistic animation or allowing users to interact with the image and create various dynamics.

English
0
1
2
113
Ugiat Technologies retweetledi
Google for Developers
Google for Developers@googledevs·
Take a close look at RAG (Retrieval Augmented Generation), an established approach used by DataGemma to generate a comprehensive response utilizing Data Commons. 🧐 → goo.gle/3XP2tl9
English
6
11
47
13.2K