Meta’s New Text-to-3D Generator Creates 3D Models in Under a Minute

A collection of detailed figurines displayed on a surface. Items include a cactus, bear, T-Rex, dragon emerging from an egg, lion, raccoon with a pizza, dogs, skull, hippo, duck, croissant, pigeon, turkey, turtle, toy car, crab, elephant, and a bird.

Whereas Meta offers with synthetic intelligence within the type of its constantly-changing content tagging system, the corporate’s analysis wing is difficult at work on novel generative AI expertise, together with a brand new Meta 3D Gen platform that delivers text-to-3D asset technology with high-quality geometry and texture.

“This method can generate 3D property with high-resolution textures & materials maps end-to-end with outcomes which can be superior in high quality to earlier state-of-the-art options — at 3-10x the pace of earlier work,” Meta AI explains on Threads.

Publish by @aiatmeta

View on Threads

Meta 3D Gen (3DGen) can create 3D property and textures from a easy textual content immediate in below a minute, per Meta’s research paper. That is functionally much like text-to-image mills like Midjourney and Adobe Firefly, however 3DGen builds absolutely 3D fashions with underlying mesh buildings that assist physically-based rendering (PBR). Which means the 3D fashions generated by Meta 3DGen can be utilized in real-world modeling and rendering purposes.

A colorful display of various 3D-printed animal figurines and objects. The collection includes lions, raccoons, a pig, birds, a cacti, and several anthropomorphic animals in costumes. Among them are also food items, beetles, and a skull, all arranged on a green surface.
It is a visible comparability of text-to-3D generations following Meta 3D Gen’s stage I (left) and stage II (proper). Per Meta, stage II generations have been most well-liked almost 70% of the time.

“Meta 3D Gen is a two-stage methodology that mixes two parts, one for text-to-3D technology and one for text-to-texture technology, respectively,” Meta explains, including that this strategy ends in “higher-quality 3D technology for immersive content material creation.”

Diagram showing stages of text-to-3D object generation. Input: "a t-rex wearing a green wool sweater" is processed into a 3D model in Stage I. In Stage II, texture is refined for the initial prompt or generated for a new prompt ("a t-rex looking like a panda"). Output: final 3D models.

3DGen combines two of Meta’s foundational generative fashions, AssetGen and TextureGen, specializing in the relative strengths of every. Meta says that primarily based on suggestions from skilled 3D artists, its new 3DGen expertise is most well-liked over competing text-to-3D fashions “a majority of the time” whereas being three to 60 occasions quicker.

A grid of 16 varied images: a beagle in a detective's outfit, a bear dressed as a lumberjack, a ceramic lion, a chihuahua in a tutu, a dachshund wearing a hat, a delicious croissant, a gold goose, a Frazer Nash Super Sport car, sourdough bread, a hippo in a sweater, a pug in a bee costume, a cow puppy, a stack of pancakes in maple syrup, a bear in medieval armor, a Mandarin duck swimming.
A curated choice of outcomes from 3DGen.

It’s value noting that by separating mesh fashions and texture maps, 3DGen guarantees important management over the ultimate output and permits for the iterative refinement frequent to text-to-image mills. Customers can alter the enter for texture model with out tweaking the underlying mannequin.

A compilation image shows various rendered objects including a stylish metal llama statue, a sushi tray with pugs, and an orc forging a hammer on an anvil. Different rendering tools and times are labeled below each object: CSM Cube 2.0, Tripo3D, Meshy v3 (refined), Rodin Gen-1, and ours.
A comparability of 3DGen outcomes (far proper column) versus competing text-to-3D fashions throughout three completely different prompts.

Meta’s complete technical paper about 3DGen goes into considerably extra element and exhibits evaluative testing outcomes in comparison with different text-to-3D fashions.


Picture credit: Meta AI

Facebook
Twitter
LinkedIn
Follow Us
Recent Posts
For Web Design