News

Introducing Stable Fast 3D: Your Gateway to Instant 3D Asset Creation from Single Images
Introducing Stable Fast 3D: Your Gateway to Instant 3D Asset Creation from Single Images
Smart Toothbrushes: Technological Innovation that Risks Your Privacy
Smart Toothbrushes: Technological Innovation that Risks Your Privacy
Global Reach with Vimeo: AI-Powered Video Translation for Any Language
Global Reach with Vimeo: AI-Powered Video Translation for Any Language
Revolutionizing Education with AI: Eureka Labs’ Mission to Transform Learning
Revolutionizing Education with AI: Eureka Labs’ Mission to Transform Learning

Artificial Intelligence

Presentan Stable Fast 3D: Tu Puerta de Entrada a la Creación Instantánea de Activos 3D a Partir de Imágenes Únicas
Presentan Stable Fast 3D: Tu Puerta de Entrada a la Creación Instantánea de Activos 3D a Partir de Imágenes Únicas

Aspectos Destacados:

  • Procesamiento Ultrarrápido: Stable Fast 3D convierte una sola imagen en un modelo 3D completo en solo 0.5 segundos.
  • Marco Mejorado: Basado en los principios de TripoSR, el modelo ahora presenta una arquitectura mejorada y funcionalidades avanzadas.
  • Aplicaciones Diversas: Ideal para desarrolladores en juegos, realidad virtual y profesionales en retail, arquitectura y diseño.
  • Acceso Conveniente: Disponible en Hugging Face bajo la Licencia de Comunidad de Stability AI, con opciones accesibles a través de Stability AI API y Stable Assistant chatbot. Empieza con una prueba gratuita para explorar y manipular tus activos 3D en Realidad Aumentada.

Estamos emocionados de presentar Stable Fast 3D, una solución avanzada de Stability AI que redefine el proceso de creación de activos 3D. Este modelo innovador transforma una imagen única en un activo 3D detallado en solo 0.5 segundos, estableciendo nuevos estándares de eficiencia y precisión en el campo de la reconstrucción 3D.

Cómo Funciona

Los usuarios comienzan cargando una imagen del objeto deseado. Luego, Stable Fast 3D genera rápidamente un modelo 3D completo, que incluye:

  • Malla UV desempaquetada
  • Atributos del material
  • Colores albedo con iluminación reducida
  • Remeshing opcional en cuads o triángulos (agregando solo 100-200 ms al tiempo total de procesamiento)

Para una explicación visual de cómo funciona este modelo y sus mejoras respecto a versiones anteriores, mira este video informativo.

La velocidad y calidad excepcionales de Stable Fast 3D lo convierten en una herramienta crucial para la creación rápida de prototipos 3D, tanto para grandes empresas como para desarrolladores independientes en gaming, realidad virtual y diseño de alto impacto.

El modelo se puede integrar fácilmente a través de Stability AI API y Stable Assistant chatbot, permitiendo visualizar e interactuar con tus modelos 3D en Realidad Aumentada en dispositivos compatibles.

Casos de Uso

Stable Fast 3D es versátil y aplicable en diversos campos, incluyendo:

  • Creación rápida de modelos durante fases de preproducción donde la experimentación es vital
  • Generación de activos estáticos para juegos (por ejemplo, objetos de fondo, muebles)
  • Creación de modelos 3D para plataformas de comercio electrónico
  • Desarrollo rápido de modelos para AR/VR

Rendimiento Inigualable

Stable Fast 3D se destaca en varias áreas clave:

  • Velocidad Excepcional: Genera activos 3D en solo 0.5 segundos en una GPU con 7GB de VRAM, o casi un segundo a través de Stability AI API.
  • Calidad Superior: Produce mallas UV desempaquetadas y parámetros de material de alta calidad, con problemas mínimos de iluminación en texturas.
  • Características Mejoradas: Incluye la generación de parámetros adicionales de material y mapas normales.

Comparado con nuestro modelo anterior SV3D, Stable Fast 3D reduce significativamente el tiempo de inferencia de 10 minutos a solo 0.5 segundos, manteniendo una calidad de salida superior.

Investigación e Innovaciones

Stable Fast 3D representa una evolución significativa respecto a nuestro modelo TripoSR anterior, incorporando un marco completamente reentrenado con modificaciones arquitectónicas avanzadas. Estas actualizaciones permiten una generación precisa de mallas y el uso de técnicas innovadoras para una rápida creación de mallas texturizadas.

Un informe técnico detallado sobre las capacidades de inferencia rápida y las innovaciones en iluminación y parámetros de material de Stable Fast 3D está disponible aquí.

Acceso y Licencias

El código del modelo Stable Fast 3D está disponible en GitHub, con pesos y opciones de demostración en Hugging Face. Está publicado bajo la Licencia de Comunidad de Stability AI, que permite el uso no comercial y el uso comercial para entidades con ingresos anuales de hasta $1M. Para organizaciones que superen este umbral de ingresos, contáctanos para opciones de licencia empresarial.

El modelo también está accesible a través de nuestra API y Stable Assistant.

Para más información, visita nuestra página del proyecto aquí.

Mantente al tanto de Stable Fast 3D y otras actualizaciones de Stability AI.

ChatGpt 4o mini
Introducing Stable Fast 3D: Your Gateway to Instant 3D Asset Creation from Single Images
Introducing Stable Fast 3D: Your Gateway to Instant 3D Asset Creation from Single Images
Global Reach with Vimeo: AI-Powered Video Translation for Any Language
Global Reach with Vimeo: AI-Powered Video Translation for Any Language
Revolutionizing Education with AI: Eureka Labs’ Mission to Transform Learning
Revolutionizing Education with AI: Eureka Labs’ Mission to Transform Learning

Programming

Sora Short Films: Understanding AI Video Pros & Cons
Sora Short Films: Understanding AI Video Pros & Cons

OpenAI’s video generation tool Sora took the AI community by storm in February with smooth, realistic videos that appear light years ahead of competitors. However, the carefully planned debut left many details out, details that have been revealed by a filmmaker who had early access to create a short using Sora.

Shy Kids is a Toronto-based digital production team that was selected by OpenAI as one of a few to produce short films, primarily for promotional purposes for OpenAI, although they were given considerable creative freedom in the creation of “Air Head.” In an interview with visual effects news outlet fxguide, post-production artist Patrick Cederberg described “the actual use of Sora” as part of his work.

Perhaps the most important lesson for most is simply this: while OpenAI’s post highlighting the short films allows the reader to assume that they more or less emerged fully formed from Sora, the reality is that these were professional productions, complete with solid storyboarding, editing, color correction and post work such as rotoscoping and visual effects. Just as Apple says “filmed with iPhone” but doesn’t show the studio setup, professional lighting, and color work after the fact, Sora’s post only talks about what it allows people to do, not how they actually did it .

Cederberg’s interview is interesting and fairly non-technical, so if you’re interested, head over to fxguide and read it. But there are some interesting details here about Sora’s use that tell us that, as impressive as it is, the model is perhaps less of a leap forward than we thought.

Control remains the most desirable and also the most elusive right now. …The closest we could get was simply being hyper-descriptive in our directions. Explaining the characters’ costumes, as well as the type of balloon, was our way of maintaining consistency because from shot to shot / generation to generation, there is not yet an established feature set to have full control over consistency.

In other words, matters that are simple in traditional filmmaking, such as choosing the color of a character’s clothing, require elaborate solutions and checks in a generative system, because each shot is created independently of the others. That could obviously change, but it’s certainly a lot more work-intensive at the moment.

Sora’s results also had to be revised to eliminate unwanted elements: Cederberg described how the model routinely generated a face on the main character’s balloon for a head, or a rope hanging down. These had to be removed in post-production, another time-consuming process, if they couldn’t get the prompt to exclude them.

Precise control of time and character or camera movements is not really possible: “There is a bit of temporal control over where these different actions occur in the actual generation, but it’s not precise… it’s more of a shot in the arm.” chance,” Cederberg said.

For example, the timing of a gesture like a greeting is a very coarse, suggestion-driven process, unlike manual animations. And a shot like a pan up on the character’s body may or may not reflect what the filmmaker wants, so the team in this case rendered a composite shot in vertical orientation and then performed a cropped pan in post-production. The generated clips were also often in slow motion for no particular reason.

In fact, the use of everyday filmmaking language, such as “pan right” or “sequence shot,” was inconsistent overall, Cederberg said, which the team found quite surprising.

“The researchers, before approaching artists to play with the tool, weren’t really thinking like filmmakers,” he said.

As a result, the team performed hundreds of generations, each lasting 10 to 20 seconds, and ended up using only a handful. Cederberg estimated the ratio at 300:1, but of course, we would probably all be surprised by the ratio in an ordinary film shoot.

The team actually made a little behind-the-scenes video explaining some of the issues they encountered, if you’re curious. Like a lot of AI-related content, the comments are quite critical of the entire effort, though not as vituperative as the AI-assisted ad we saw criticized recently.

The last interesting complication concerns copyright: If you ask Sora to give you a clip from “Star Wars,” he will refuse. And if you try to get around it with “robed man with lightsaber on retro-futuristic spaceship”, it will refuse too, since by some mechanism it recognizes what you’re trying to do. He also refused to do an “Aronofsky-type” shot or a “Hitchcock zoom.”

On the one hand, it makes sense. But it begs the question: If Sora knows what these things are, does that mean the model was trained with that content, to better recognize what’s infringing? OpenAI, which keeps its training data cards close to its vest, to the point of absurdity, as with CTO Mira Murati’s interview with Joanna Stern, will almost certainly never tell us.

As for Sora and his use in filmmaking, he’s clearly a powerful and useful tool in his place, but his place is not “creating movies out of thin air.” Still. As another villain once famously said, “that will come later.”

Enlace al artículo sobre Elon Musk
Exploring Angular 17: Unveiling the Latest Innovations
Exploring Angular 17: Unveiling the Latest Innovations
The incredible world of technological innovations in Angular
The incredible world of technological innovations in Angular
Exploring the Future: The Programming Frameworks that Revolutionize 2023
Exploring the Future: The Programming Frameworks that Revolutionize 2023

Latest News

Verified by MonsterInsights