AI Engineer: Generative Geometry for Hardware Design
Vinci4D.ai
Location
Palo Alto HQ
Employment Type
Full time
Location Type
Hybrid
Department
Engineering
About Us
We're building a co-pilot for hardware designers. Our mission is to enable millions of hardware designers and engineers to iterate through designs 1000x faster.
We are building our geometry + physics driven foundation model for each class of part design. Our first model is shipped and we are expanding our capabilities!
Backed by Khosla Ventures and Eclipse Ventures
About You
You’ve shipped AI products that operate in high-dimensional, multimodal domains — computer vision, geometry, or simulation-based workflows. You have experience building models that don’t just analyze data, but generate complex, structured outputs under real-world constraints.
You’re comfortable navigating both classic modeling techniques and modern deep learning architectures, and you care about building systems that are principled, testable, and physically meaningful.
What You’ll Work On
Design conditional generative models for 3D geometry tailored to hardware design workflows, including mesh-based, parametric (e.g., CAD) and implicit representations
Develop models that generate geometry conditioned on constraints, partial designs, simulation outcomes, or functional requirements.
Support inverse design tasks where the model proposes viable geometries given desired performance or physical behavior
Implement cutting-edge generative architectures for 3D data such as:
Diffusion models for point clouds, voxel grids, or triangle meshes
Neural implicit representations (SDFs, DeepSDF, NeRF variants for shape modeling)
Transformer or autoregressive models for topological and geometric sequence modeling
CAD-aware generation pipelines (sketch-based or parametric component generators)
Develop pipelines for geometry-aware learning and generation combining:
Mesh and geometry processing (remeshing, simplification, subdivision)
Differentiable simulation or physics-informed learning components
Conditioning on design constraints, performance targets, or class-specific priors
Collaborate with domain experts in physics, geometry, and simulation to:
Integrate physical principles and simulation feedback into the generation loop
Ensure designs meet functional, physical, and manufacturability requirements
Translate domain knowledge into data priors, architectural biases, or constraints
Design experiments and benchmarks to evaluate generation quality such as:
Geometry fidelity and resolution
Physical plausibility and constraint satisfaction
Generalization to novel design tasks or unseen part types
Build product-facing generative tools, including:
Auto-complete or correction of partial designs
LLM to CAD generation
Proposal of high-quality geometry variants from a design prompt
Design-space exploration tools guided by downstream simulation outcomes
Own projects end-to-end: rapidly prototype models, test ideas, gather feedback and contribute to production deployment in collaboration with cross-functional teams
Qualifications
4+ years of experience developing and shipping products
Strong background in deep learning, especially applied to 3D or spatial data.
Hands-on experience with mesh generation, implicit surfaces, or neural fields (e.g., NeRF, SDF, DeepSDF, Occupancy Networks).
Experience with the related technologies, libraries, and languages: Python, C++, PyTorch (3D)/TensorFlow/JAX; plus to have GPU programming
Experience with diffusion models for 3D generation
Startup experience is a strong advantage.
Understanding of geometry representations (mesh, voxel, point cloud, NURBS, parametric surfaces).
Familiarity with 3D geometry processing, including mesh handling, surface reconstruction, spatial data structures, and basic topology, to support effective 3D model manipulation and analysis